[ 542.123216] env[62952]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62952) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 542.123551] env[62952]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62952) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 542.123679] env[62952]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62952) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 542.123978] env[62952]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 542.219460] env[62952]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62952) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 542.229451] env[62952]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62952) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 542.830377] env[62952]: INFO nova.virt.driver [None req-2b1defb8-564c-4442-8411-d12bb27d5b03 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 542.918911] env[62952]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.919090] env[62952]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.919203] env[62952]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62952) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 546.017156] env[62952]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-31536c3f-2b43-40cc-a6ab-6e2e8d2a6211 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.034027] env[62952]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62952) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 546.034027] env[62952]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-f1606a17-3690-4a52-a153-4b8cdd31a9ff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.066641] env[62952]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 28389. [ 546.066641] env[62952]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.147s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.066641] env[62952]: INFO nova.virt.vmwareapi.driver [None req-2b1defb8-564c-4442-8411-d12bb27d5b03 None None] VMware vCenter version: 7.0.3 [ 546.069828] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88adf60c-f67f-4de8-8dd0-c2a39a2e6f88 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.092301] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-032e7c8c-6bbd-4d03-91d3-9edfaeff65e8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.095019] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19647039-42a2-4c9f-a764-b174c86cc7f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.101840] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28bd4958-ee9b-415c-8ce7-940906138ae9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.115012] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66508ab6-0b01-41f7-8501-92b1dff25d74 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.120874] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e70731-f353-47d3-a75a-b453a1719679 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.150515] env[62952]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-767bb491-56f8-4f51-b85c-05b7b36e7d6d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.155453] env[62952]: DEBUG nova.virt.vmwareapi.driver [None req-2b1defb8-564c-4442-8411-d12bb27d5b03 None None] Extension org.openstack.compute already exists. {{(pid=62952) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 546.158160] env[62952]: INFO nova.compute.provider_config [None req-2b1defb8-564c-4442-8411-d12bb27d5b03 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 546.662066] env[62952]: DEBUG nova.context [None req-2b1defb8-564c-4442-8411-d12bb27d5b03 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),82ed63a6-d57a-48ea-a594-ed5abf3e72dc(cell1) {{(pid=62952) load_cells /opt/stack/nova/nova/context.py:464}} [ 546.665100] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.665329] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.666013] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.666442] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] Acquiring lock "82ed63a6-d57a-48ea-a594-ed5abf3e72dc" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.666642] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] Lock "82ed63a6-d57a-48ea-a594-ed5abf3e72dc" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.667651] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] Lock "82ed63a6-d57a-48ea-a594-ed5abf3e72dc" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.687366] env[62952]: INFO dbcounter [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] Registered counter for database nova_cell0 [ 546.695273] env[62952]: INFO dbcounter [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] Registered counter for database nova_cell1 [ 546.698437] env[62952]: DEBUG oslo_db.sqlalchemy.engines [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62952) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 546.698789] env[62952]: DEBUG oslo_db.sqlalchemy.engines [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62952) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 546.703672] env[62952]: ERROR nova.db.main.api [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.703672] env[62952]: result = function(*args, **kwargs) [ 546.703672] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 546.703672] env[62952]: return func(*args, **kwargs) [ 546.703672] env[62952]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 546.703672] env[62952]: result = fn(*args, **kwargs) [ 546.703672] env[62952]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 546.703672] env[62952]: return f(*args, **kwargs) [ 546.703672] env[62952]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 546.703672] env[62952]: return db.service_get_minimum_version(context, binaries) [ 546.703672] env[62952]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 546.703672] env[62952]: _check_db_access() [ 546.703672] env[62952]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 546.703672] env[62952]: stacktrace = ''.join(traceback.format_stack()) [ 546.703672] env[62952]: [ 546.704411] env[62952]: ERROR nova.db.main.api [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.704411] env[62952]: result = function(*args, **kwargs) [ 546.704411] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 546.704411] env[62952]: return func(*args, **kwargs) [ 546.704411] env[62952]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 546.704411] env[62952]: result = fn(*args, **kwargs) [ 546.704411] env[62952]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 546.704411] env[62952]: return f(*args, **kwargs) [ 546.704411] env[62952]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 546.704411] env[62952]: return db.service_get_minimum_version(context, binaries) [ 546.704411] env[62952]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 546.704411] env[62952]: _check_db_access() [ 546.704411] env[62952]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 546.704411] env[62952]: stacktrace = ''.join(traceback.format_stack()) [ 546.704411] env[62952]: [ 546.704833] env[62952]: WARNING nova.objects.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 546.704929] env[62952]: WARNING nova.objects.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] Failed to get minimum service version for cell 82ed63a6-d57a-48ea-a594-ed5abf3e72dc [ 546.705403] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] Acquiring lock "singleton_lock" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.705614] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] Acquired lock "singleton_lock" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.705896] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] Releasing lock "singleton_lock" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.706308] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] Full set of CONF: {{(pid=62952) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 546.706453] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ******************************************************************************** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 546.706582] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] Configuration options gathered from: {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 546.706716] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 546.706903] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 546.707039] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ================================================================================ {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 546.707253] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] allow_resize_to_same_host = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.707426] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] arq_binding_timeout = 300 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.707558] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] backdoor_port = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.707684] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] backdoor_socket = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.707848] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] block_device_allocate_retries = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.708014] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] block_device_allocate_retries_interval = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.708187] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cert = self.pem {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.708353] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.708521] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] compute_monitors = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.708687] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] config_dir = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.708854] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] config_drive_format = iso9660 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.708986] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.709162] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] config_source = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.709329] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] console_host = devstack {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.709490] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] control_exchange = nova {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.709647] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cpu_allocation_ratio = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.709804] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] daemon = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.709966] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] debug = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.710133] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] default_access_ip_network_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.710298] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] default_availability_zone = nova {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.710451] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] default_ephemeral_format = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.710608] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] default_green_pool_size = 1000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.710843] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.711025] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] default_schedule_zone = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.711187] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] disk_allocation_ratio = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.711345] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] enable_new_services = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.711517] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] enabled_apis = ['osapi_compute'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.711682] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] enabled_ssl_apis = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.711845] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] flat_injected = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.712008] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] force_config_drive = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.712176] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] force_raw_images = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.712346] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] graceful_shutdown_timeout = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.712501] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] heal_instance_info_cache_interval = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.712708] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] host = cpu-1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.712876] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.713049] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] initial_disk_allocation_ratio = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.713231] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] initial_ram_allocation_ratio = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.713456] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.713621] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] instance_build_timeout = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.713781] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] instance_delete_interval = 300 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.713944] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] instance_format = [instance: %(uuid)s] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.714125] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] instance_name_template = instance-%08x {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.714327] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] instance_usage_audit = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.714476] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] instance_usage_audit_period = month {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.714618] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.714785] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] instances_path = /opt/stack/data/nova/instances {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.714951] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] internal_service_availability_zone = internal {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.715126] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] key = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.715315] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] live_migration_retry_count = 30 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.715523] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] log_color = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.715672] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] log_config_append = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.715840] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.715998] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] log_dir = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.716173] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] log_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.716302] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] log_options = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.716467] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] log_rotate_interval = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.716632] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] log_rotate_interval_type = days {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.716794] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] log_rotation_type = none {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.716921] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.717055] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.717223] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.717387] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.717513] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.717671] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] long_rpc_timeout = 1800 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.717829] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] max_concurrent_builds = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.717986] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] max_concurrent_live_migrations = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.718154] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] max_concurrent_snapshots = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.718310] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] max_local_block_devices = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.718465] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] max_logfile_count = 30 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.718621] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] max_logfile_size_mb = 200 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.718782] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] maximum_instance_delete_attempts = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.718948] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] metadata_listen = 0.0.0.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.719124] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] metadata_listen_port = 8775 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.719293] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] metadata_workers = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.719460] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] migrate_max_retries = -1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.719623] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] mkisofs_cmd = genisoimage {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.719824] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] my_block_storage_ip = 10.180.1.21 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.719955] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] my_ip = 10.180.1.21 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.720127] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] network_allocate_retries = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.720314] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.720486] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] osapi_compute_listen = 0.0.0.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.720650] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] osapi_compute_listen_port = 8774 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.720818] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] osapi_compute_unique_server_name_scope = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.720982] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] osapi_compute_workers = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.721159] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] password_length = 12 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.721320] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] periodic_enable = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.721479] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] periodic_fuzzy_delay = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.721642] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] pointer_model = usbtablet {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.721804] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] preallocate_images = none {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.721961] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] publish_errors = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.722098] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] pybasedir = /opt/stack/nova {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.722257] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ram_allocation_ratio = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.722419] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] rate_limit_burst = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.722586] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] rate_limit_except_level = CRITICAL {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.722744] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] rate_limit_interval = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.722897] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] reboot_timeout = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.723063] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] reclaim_instance_interval = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.723248] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] record = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.723432] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] reimage_timeout_per_gb = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.723597] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] report_interval = 120 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.723757] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] rescue_timeout = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.723915] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] reserved_host_cpus = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.724083] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] reserved_host_disk_mb = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.724262] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] reserved_host_memory_mb = 512 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.724452] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] reserved_huge_pages = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.724619] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] resize_confirm_window = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.724778] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] resize_fs_using_block_device = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.724935] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] resume_guests_state_on_host_boot = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.725121] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.725287] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] rpc_response_timeout = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.725448] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] run_external_periodic_tasks = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.725644] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] running_deleted_instance_action = reap {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.725813] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] running_deleted_instance_poll_interval = 1800 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.725973] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] running_deleted_instance_timeout = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.726152] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] scheduler_instance_sync_interval = 120 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.726324] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] service_down_time = 720 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.726489] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] servicegroup_driver = db {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.726642] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] shell_completion = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.726801] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] shelved_offload_time = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.726958] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] shelved_poll_interval = 3600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.727134] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] shutdown_timeout = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.727295] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] source_is_ipv6 = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.727453] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ssl_only = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.727693] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.727859] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] sync_power_state_interval = 600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.728028] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] sync_power_state_pool_size = 1000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.728198] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] syslog_log_facility = LOG_USER {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.728357] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] tempdir = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.728513] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] timeout_nbd = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.728676] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] transport_url = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.728836] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] update_resources_interval = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.728995] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] use_cow_images = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.729691] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] use_eventlog = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.729879] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] use_journal = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.730061] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] use_json = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.730233] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] use_rootwrap_daemon = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.730397] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] use_stderr = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.730560] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] use_syslog = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.730758] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vcpu_pin_set = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.730935] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vif_plugging_is_fatal = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.731119] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vif_plugging_timeout = 300 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.731287] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] virt_mkfs = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.731451] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] volume_usage_poll_interval = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.731613] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] watch_log_file = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.731780] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] web = /usr/share/spice-html5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 546.731964] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_concurrency.disable_process_locking = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.732262] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.732445] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.732612] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.732782] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.732951] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.733129] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.733337] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.auth_strategy = keystone {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.733505] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.compute_link_prefix = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.733683] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.733857] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.dhcp_domain = novalocal {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.734041] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.enable_instance_password = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.734212] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.glance_link_prefix = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.734400] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.734581] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.734749] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.instance_list_per_project_cells = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.734914] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.list_records_by_skipping_down_cells = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.735089] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.local_metadata_per_cell = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.735263] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.max_limit = 1000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.735430] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.metadata_cache_expiration = 15 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.735608] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.neutron_default_tenant_id = default {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.735773] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.response_validation = warn {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.735941] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.use_neutron_default_nets = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.736120] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.736285] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.736459] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.736629] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.736800] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.vendordata_dynamic_targets = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.736963] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.vendordata_jsonfile_path = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.737159] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.737356] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.backend = dogpile.cache.memcached {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.737526] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.backend_argument = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.737697] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.config_prefix = cache.oslo {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.737867] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.dead_timeout = 60.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.738042] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.debug_cache_backend = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.738210] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.enable_retry_client = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.738374] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.enable_socket_keepalive = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.738546] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.enabled = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.738707] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.enforce_fips_mode = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.738870] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.expiration_time = 600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.739043] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.hashclient_retry_attempts = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.739212] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.hashclient_retry_delay = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.739378] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.memcache_dead_retry = 300 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.739536] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.memcache_password = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.739700] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.739863] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.740034] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.memcache_pool_maxsize = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.740205] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.740369] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.memcache_sasl_enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.740544] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.740737] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.memcache_socket_timeout = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.740875] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.memcache_username = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.741051] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.proxies = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.741219] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.redis_db = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.741379] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.redis_password = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.741548] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.redis_sentinel_service_name = mymaster {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.741722] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.741890] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.redis_server = localhost:6379 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.742072] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.redis_socket_timeout = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.742235] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.redis_username = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.742400] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.retry_attempts = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.742564] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.retry_delay = 0.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.742727] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.socket_keepalive_count = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.742914] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.socket_keepalive_idle = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.743061] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.socket_keepalive_interval = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.743275] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.tls_allowed_ciphers = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.743454] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.tls_cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.743618] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.tls_certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.743780] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.tls_enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.743938] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cache.tls_keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.744140] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cinder.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.744414] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cinder.auth_type = password {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.744614] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cinder.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.744775] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cinder.catalog_info = volumev3::publicURL {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.744933] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cinder.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.745110] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cinder.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.745276] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cinder.cross_az_attach = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.745440] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cinder.debug = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.745600] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cinder.endpoint_template = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.745763] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cinder.http_retries = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.745923] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cinder.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.746091] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cinder.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.746267] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cinder.os_region_name = RegionOne {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.746432] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cinder.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.746590] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cinder.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.746761] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.746920] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] compute.cpu_dedicated_set = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.747089] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] compute.cpu_shared_set = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.747256] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] compute.image_type_exclude_list = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.747425] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.747588] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] compute.max_concurrent_disk_ops = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.747747] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] compute.max_disk_devices_to_attach = -1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.747910] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.748089] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.748257] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] compute.resource_provider_association_refresh = 300 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.748422] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.748583] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] compute.shutdown_retry_interval = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.748763] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.748940] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] conductor.workers = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.749130] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] console.allowed_origins = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.749293] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] console.ssl_ciphers = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.749464] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] console.ssl_minimum_version = default {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.749634] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] consoleauth.enforce_session_timeout = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.749802] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] consoleauth.token_ttl = 600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.749975] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.750146] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.750313] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.750475] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.connect_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.750632] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.connect_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.750789] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.endpoint_override = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.750952] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.751122] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.751283] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.max_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.751441] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.min_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.751596] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.region_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.751752] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.retriable_status_codes = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.751908] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.service_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.752083] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.service_type = accelerator {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.752247] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.752407] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.status_code_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.752563] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.status_code_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.752719] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.752899] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.753072] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] cyborg.version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.753277] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.backend = sqlalchemy {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.753460] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.connection = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.753627] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.connection_debug = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.753793] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.connection_parameters = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.753956] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.connection_recycle_time = 3600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.754157] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.connection_trace = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.754387] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.db_inc_retry_interval = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.754633] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.db_max_retries = 20 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.754864] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.db_max_retry_interval = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.755093] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.db_retry_interval = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.755319] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.max_overflow = 50 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.755548] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.max_pool_size = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.755760] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.max_retries = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.755985] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.756227] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.mysql_wsrep_sync_wait = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.756435] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.pool_timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.756660] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.retry_interval = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.756877] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.slave_connection = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.757108] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.sqlite_synchronous = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.757337] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] database.use_db_reconnect = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.757585] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.backend = sqlalchemy {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.757791] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.connection = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.757962] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.connection_debug = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.758150] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.connection_parameters = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.758318] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.connection_recycle_time = 3600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.758485] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.connection_trace = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.758647] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.db_inc_retry_interval = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.758812] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.db_max_retries = 20 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.758973] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.db_max_retry_interval = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.759149] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.db_retry_interval = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.759316] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.max_overflow = 50 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.759480] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.max_pool_size = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.759642] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.max_retries = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.759810] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.759969] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.760140] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.pool_timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.760306] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.retry_interval = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.760469] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.slave_connection = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.760632] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] api_database.sqlite_synchronous = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.760805] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] devices.enabled_mdev_types = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.760981] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.761167] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ephemeral_storage_encryption.default_format = luks {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.761332] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ephemeral_storage_encryption.enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.761497] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.761668] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.api_servers = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.761833] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.761993] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.762183] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.762347] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.connect_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.762507] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.connect_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.762668] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.debug = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.762830] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.default_trusted_certificate_ids = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.762992] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.enable_certificate_validation = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.763175] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.enable_rbd_download = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.763349] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.endpoint_override = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.763521] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.763683] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.763841] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.max_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.763998] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.min_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.764176] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.num_retries = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.764362] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.rbd_ceph_conf = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.764536] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.rbd_connect_timeout = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.764706] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.rbd_pool = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.764872] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.rbd_user = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.765055] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.region_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.765200] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.retriable_status_codes = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.765362] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.service_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.765531] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.service_type = image {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.765692] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.765851] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.status_code_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.766013] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.status_code_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.766178] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.766368] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.766533] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.verify_glance_signatures = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.766695] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] glance.version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.766864] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] guestfs.debug = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.767040] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] mks.enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.767393] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.767586] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] image_cache.manager_interval = 2400 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.767758] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] image_cache.precache_concurrency = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.767928] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] image_cache.remove_unused_base_images = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.768110] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.768282] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.768459] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] image_cache.subdirectory_name = _base {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.768634] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.api_max_retries = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.768800] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.api_retry_interval = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.768960] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.769139] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.auth_type = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.769300] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.769460] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.769623] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.769787] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.conductor_group = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.769946] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.connect_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.770115] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.connect_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.770274] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.endpoint_override = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.770441] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.770598] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.770756] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.max_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.770914] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.min_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.771092] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.peer_list = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.771257] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.region_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.771418] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.retriable_status_codes = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.771581] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.serial_console_state_timeout = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.771741] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.service_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.771910] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.service_type = baremetal {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.772080] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.shard = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.772249] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.772412] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.status_code_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.772567] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.status_code_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.772725] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.772907] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.773078] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ironic.version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.773281] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.773465] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] key_manager.fixed_key = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.773648] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.773811] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican.barbican_api_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.773973] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican.barbican_endpoint = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.774157] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican.barbican_endpoint_type = public {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.774337] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican.barbican_region_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.774513] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.774674] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.774838] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.775018] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.775180] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.775347] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican.number_of_retries = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.775511] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican.retry_delay = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.775673] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican.send_service_user_token = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.775837] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.775996] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.776180] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican.verify_ssl = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.776342] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican.verify_ssl_path = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.776508] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican_service_user.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.776671] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican_service_user.auth_type = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.776831] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican_service_user.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.776988] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican_service_user.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.777165] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican_service_user.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.777327] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican_service_user.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.777488] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican_service_user.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.777650] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican_service_user.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.777810] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] barbican_service_user.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.777974] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vault.approle_role_id = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.778143] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vault.approle_secret_id = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.778313] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vault.kv_mountpoint = secret {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.778475] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vault.kv_path = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.778639] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vault.kv_version = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.778797] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vault.namespace = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.778957] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vault.root_token_id = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.779127] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vault.ssl_ca_crt_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.779297] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vault.timeout = 60.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.779463] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vault.use_ssl = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.779632] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.779806] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.779967] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.auth_type = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.780139] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.780298] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.780463] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.780621] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.connect_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.780780] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.connect_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.780937] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.endpoint_override = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.781109] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.781268] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.781426] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.max_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.781582] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.min_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.781740] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.region_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.781897] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.retriable_status_codes = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.782064] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.service_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.782234] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.service_type = identity {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.782400] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.782558] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.status_code_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.782717] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.status_code_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.782874] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.783064] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.783252] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] keystone.version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.783457] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.connection_uri = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.783620] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.cpu_mode = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.783788] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.cpu_model_extra_flags = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.783956] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.cpu_models = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.784141] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.cpu_power_governor_high = performance {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.784329] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.cpu_power_governor_low = powersave {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.784506] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.cpu_power_management = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.784678] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.784842] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.device_detach_attempts = 8 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.785035] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.device_detach_timeout = 20 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.785218] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.disk_cachemodes = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.785370] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.disk_prefix = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.785542] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.enabled_perf_events = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.785705] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.file_backed_memory = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.785869] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.gid_maps = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.786038] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.hw_disk_discard = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.786200] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.hw_machine_type = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.786369] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.images_rbd_ceph_conf = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.786594] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.786867] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.787175] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.images_rbd_glance_store_name = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.787481] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.images_rbd_pool = rbd {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.787774] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.images_type = default {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.788025] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.images_volume_group = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.788220] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.inject_key = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.788396] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.inject_partition = -2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.788564] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.inject_password = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.788729] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.iscsi_iface = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.788893] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.iser_use_multipath = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.789070] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.live_migration_bandwidth = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.789241] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.789408] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.live_migration_downtime = 500 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.789573] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.789735] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.789896] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.live_migration_inbound_addr = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.790068] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.790234] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.live_migration_permit_post_copy = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.790396] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.live_migration_scheme = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.790568] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.live_migration_timeout_action = abort {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.790734] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.live_migration_tunnelled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.790899] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.live_migration_uri = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.791078] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.live_migration_with_native_tls = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.791273] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.max_queues = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.791460] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.791690] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.791853] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.nfs_mount_options = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.793054] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.793271] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.793463] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.num_iser_scan_tries = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.793635] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.num_memory_encrypted_guests = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.793805] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.793974] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.num_pcie_ports = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.794156] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.num_volume_scan_tries = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.794326] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.pmem_namespaces = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.794490] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.quobyte_client_cfg = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.794777] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.794951] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.rbd_connect_timeout = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.795132] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.795307] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.795514] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.rbd_secret_uuid = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.795692] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.rbd_user = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.795858] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.796048] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.remote_filesystem_transport = ssh {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.796219] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.rescue_image_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.796382] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.rescue_kernel_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.796544] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.rescue_ramdisk_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.796715] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.796877] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.rx_queue_size = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.797059] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.smbfs_mount_options = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.797345] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.797524] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.snapshot_compression = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.797688] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.snapshot_image_format = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.797909] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.798089] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.sparse_logical_volumes = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.798258] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.swtpm_enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.798433] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.swtpm_group = tss {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.798602] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.swtpm_user = tss {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.798774] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.sysinfo_serial = unique {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.798934] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.tb_cache_size = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.799105] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.tx_queue_size = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.799273] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.uid_maps = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.799438] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.use_virtio_for_bridges = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.799607] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.virt_type = kvm {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.799776] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.volume_clear = zero {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.799938] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.volume_clear_size = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.800113] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.volume_use_multipath = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.800275] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.vzstorage_cache_path = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.800443] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.800606] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.vzstorage_mount_group = qemu {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.800767] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.vzstorage_mount_opts = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.800933] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.801235] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.801422] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.vzstorage_mount_user = stack {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.801588] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.801760] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.801932] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.auth_type = password {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.802104] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.802268] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.802433] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.802625] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.connect_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.802784] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.connect_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.802952] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.default_floating_pool = public {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.803119] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.endpoint_override = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.803335] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.extension_sync_interval = 600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.803519] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.http_retries = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.803684] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.803842] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.804012] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.max_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.804189] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.804346] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.min_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.804511] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.ovs_bridge = br-int {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.804674] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.physnets = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.804839] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.region_name = RegionOne {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.804996] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.retriable_status_codes = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.805181] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.service_metadata_proxy = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.805336] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.service_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.805524] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.service_type = network {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.805663] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.805817] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.status_code_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.805973] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.status_code_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.806142] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.806350] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.806513] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] neutron.version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.806733] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] notifications.bdms_in_notifications = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.806914] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] notifications.default_level = INFO {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.807099] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] notifications.notification_format = unversioned {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.807265] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] notifications.notify_on_state_change = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.807442] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.807618] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] pci.alias = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.807787] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] pci.device_spec = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.807955] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] pci.report_in_placement = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.808140] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.808318] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.auth_type = password {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.808487] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.808649] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.808806] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.808967] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.809137] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.connect_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.809299] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.connect_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.809460] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.default_domain_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.809616] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.default_domain_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.809774] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.domain_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.809930] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.domain_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.810097] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.endpoint_override = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.810261] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.810419] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.810575] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.max_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.810732] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.min_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.810898] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.password = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.811067] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.project_domain_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.811254] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.project_domain_name = Default {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.811445] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.project_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.811623] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.project_name = service {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.811791] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.region_name = RegionOne {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.811954] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.retriable_status_codes = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.812129] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.service_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.812303] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.service_type = placement {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.812469] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.812629] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.status_code_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.812793] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.status_code_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.812974] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.system_scope = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.813122] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.813305] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.trust_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.813473] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.user_domain_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.813644] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.user_domain_name = Default {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.813804] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.user_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.813977] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.username = nova {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.814174] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.814338] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] placement.version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.814516] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] quota.cores = 20 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.814682] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] quota.count_usage_from_placement = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.814854] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.815033] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] quota.injected_file_content_bytes = 10240 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.815207] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] quota.injected_file_path_length = 255 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.815408] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] quota.injected_files = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.815628] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] quota.instances = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.815754] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] quota.key_pairs = 100 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.815920] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] quota.metadata_items = 128 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.816099] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] quota.ram = 51200 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.816294] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] quota.recheck_quota = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.816471] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] quota.server_group_members = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.816640] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] quota.server_groups = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.816808] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.816973] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.817148] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] scheduler.image_metadata_prefilter = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.817310] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.817475] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] scheduler.max_attempts = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.817637] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] scheduler.max_placement_results = 1000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.817802] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.817964] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] scheduler.query_placement_for_image_type_support = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.818141] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.818319] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] scheduler.workers = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.818500] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.818669] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.818846] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.819020] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.819187] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.819354] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.819517] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.819704] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.819871] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.host_subset_size = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.820054] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.820222] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.820386] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.820551] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.isolated_hosts = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.820725] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.isolated_images = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.820889] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.821061] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.821246] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.821425] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.pci_in_placement = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.821592] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.821754] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.821916] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.822089] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.822256] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.822420] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.822582] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.track_instance_changes = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.822757] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.822926] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] metrics.required = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.823101] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] metrics.weight_multiplier = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.823297] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.823467] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] metrics.weight_setting = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.823794] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.823971] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] serial_console.enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.824168] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] serial_console.port_range = 10000:20000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.824398] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.824578] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.824748] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] serial_console.serialproxy_port = 6083 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.824918] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] service_user.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.825107] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] service_user.auth_type = password {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.825270] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] service_user.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.825453] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] service_user.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.825609] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] service_user.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.825741] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] service_user.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.825897] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] service_user.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.826077] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] service_user.send_service_user_token = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.826244] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] service_user.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.826404] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] service_user.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.826571] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] spice.agent_enabled = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.826730] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] spice.enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.827067] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.827264] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.827471] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] spice.html5proxy_port = 6082 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.827642] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] spice.image_compression = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.827802] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] spice.jpeg_compression = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.827962] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] spice.playback_compression = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.828146] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] spice.server_listen = 127.0.0.1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.828320] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.828483] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] spice.streaming_mode = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.828641] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] spice.zlib_compression = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.828807] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] upgrade_levels.baseapi = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.828976] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] upgrade_levels.compute = auto {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.829149] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] upgrade_levels.conductor = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.829309] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] upgrade_levels.scheduler = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.829477] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vendordata_dynamic_auth.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.829640] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vendordata_dynamic_auth.auth_type = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.829797] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vendordata_dynamic_auth.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.829955] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vendordata_dynamic_auth.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.830126] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.830286] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vendordata_dynamic_auth.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.830445] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vendordata_dynamic_auth.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.830605] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.830764] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vendordata_dynamic_auth.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.830935] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.api_retry_count = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.831106] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.ca_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.831300] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.cache_prefix = devstack-image-cache {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.831474] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.cluster_name = testcl1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.831639] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.connection_pool_size = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.831799] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.console_delay_seconds = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.831968] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.datastore_regex = ^datastore.* {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.832192] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.832368] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.host_password = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.832536] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.host_port = 443 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.832704] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.host_username = administrator@vsphere.local {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.832873] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.insecure = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.833044] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.integration_bridge = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.833235] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.maximum_objects = 100 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.833407] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.pbm_default_policy = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.833572] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.pbm_enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.833729] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.pbm_wsdl_location = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.833898] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.834068] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.serial_port_proxy_uri = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.834230] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.serial_port_service_uri = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.834397] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.task_poll_interval = 0.5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.834568] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.use_linked_clone = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.834735] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.vnc_keymap = en-us {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.834898] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.vnc_port = 5900 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.835071] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vmware.vnc_port_total = 10000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.835259] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vnc.auth_schemes = ['none'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.835434] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vnc.enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.835738] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.835923] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.836109] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vnc.novncproxy_port = 6080 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.836291] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vnc.server_listen = 127.0.0.1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.836468] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.836630] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vnc.vencrypt_ca_certs = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.836790] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vnc.vencrypt_client_cert = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.836951] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vnc.vencrypt_client_key = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.837138] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.837304] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.disable_deep_image_inspection = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.837471] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.837631] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.837792] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.837952] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.disable_rootwrap = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.838122] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.enable_numa_live_migration = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.838285] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.838445] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.838605] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.838762] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.libvirt_disable_apic = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.838923] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.839093] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.839255] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.839444] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.839627] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.839790] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.839951] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.840125] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.840289] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.840457] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.840640] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.840807] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] wsgi.client_socket_timeout = 900 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.840974] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] wsgi.default_pool_size = 1000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.841157] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] wsgi.keep_alive = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.841325] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] wsgi.max_header_line = 16384 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.841488] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] wsgi.secure_proxy_ssl_header = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.841647] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] wsgi.ssl_ca_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.841807] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] wsgi.ssl_cert_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.841966] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] wsgi.ssl_key_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.842142] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] wsgi.tcp_keepidle = 600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.842362] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.842562] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] zvm.ca_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.842728] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] zvm.cloud_connector_url = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.843021] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.843220] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] zvm.reachable_timeout = 300 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.843411] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_policy.enforce_new_defaults = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.843781] env[62952]: WARNING oslo_config.cfg [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 546.843969] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_policy.enforce_scope = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.844184] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_policy.policy_default_rule = default {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.844378] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.844554] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_policy.policy_file = policy.yaml {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.844749] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.844923] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.845096] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.845259] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.845425] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.845600] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.845789] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.845984] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] profiler.connection_string = messaging:// {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.846120] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] profiler.enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.846290] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] profiler.es_doc_type = notification {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.846457] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] profiler.es_scroll_size = 10000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.846627] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] profiler.es_scroll_time = 2m {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.846789] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] profiler.filter_error_trace = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.846957] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] profiler.hmac_keys = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.847145] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] profiler.sentinel_service_name = mymaster {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.847315] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] profiler.socket_timeout = 0.1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.847482] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] profiler.trace_requests = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.847643] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] profiler.trace_sqlalchemy = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.847815] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] profiler_jaeger.process_tags = {} {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.847974] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] profiler_jaeger.service_name_prefix = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.848150] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] profiler_otlp.service_name_prefix = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.848317] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] remote_debug.host = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.848478] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] remote_debug.port = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.848656] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.848823] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.848984] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.849158] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.849322] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.849485] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.849646] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.849808] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.849969] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.850152] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.850314] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.850485] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.850652] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.850820] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.850988] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.851175] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.851373] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.851555] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.851720] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.851882] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.852057] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.852228] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.852394] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.852560] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.852721] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.852884] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.853058] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.853249] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.853430] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.853597] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.ssl = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.853769] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.853937] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.854132] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.854335] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.854511] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.ssl_version = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.854674] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.854861] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.855038] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_notifications.retry = -1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.855225] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.855401] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_messaging_notifications.transport_url = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.855571] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.855739] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.auth_type = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.855894] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.856071] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.856232] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.856432] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.connect_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.856609] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.connect_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.856774] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.endpoint_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.856935] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.endpoint_override = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.857112] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.857274] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.857435] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.max_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.857593] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.min_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.857750] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.region_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.857912] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.retriable_status_codes = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.858082] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.service_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.858266] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.service_type = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.858458] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.858621] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.status_code_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.858782] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.status_code_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.858940] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.859110] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.valid_interfaces = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.859270] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_limit.version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.859445] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_reports.file_event_handler = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.859605] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.859762] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] oslo_reports.log_dir = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.859931] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.860103] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.860265] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.860431] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.860595] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.860751] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.860920] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.861089] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vif_plug_ovs_privileged.group = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.861273] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.861449] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.861613] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.861771] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] vif_plug_ovs_privileged.user = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.861939] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_vif_linux_bridge.flat_interface = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.862143] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.862310] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.862482] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.862649] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.862812] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.862975] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.863150] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.863383] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.863571] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_vif_ovs.isolate_vif = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.863739] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.863906] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.864087] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.864261] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_vif_ovs.ovsdb_interface = native {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.864425] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_vif_ovs.per_port_bridge = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.864591] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_brick.lock_path = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.864755] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.864918] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.865100] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] privsep_osbrick.capabilities = [21] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.865264] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] privsep_osbrick.group = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.865423] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] privsep_osbrick.helper_command = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.865588] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.865750] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.865961] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] privsep_osbrick.user = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.866088] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.866251] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] nova_sys_admin.group = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.866409] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] nova_sys_admin.helper_command = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.866571] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.866732] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.866887] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] nova_sys_admin.user = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 546.867024] env[62952]: DEBUG oslo_service.service [None req-d963e7df-e1ab-46b2-9e56-33b94bc00f98 None None] ******************************************************************************** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 546.867500] env[62952]: INFO nova.service [-] Starting compute node (version 0.1.0) [ 547.370716] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Getting list of instances from cluster (obj){ [ 547.370716] env[62952]: value = "domain-c8" [ 547.370716] env[62952]: _type = "ClusterComputeResource" [ 547.370716] env[62952]: } {{(pid=62952) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 547.372052] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c5f1c08-2e0a-46c7-a150-2b9109cbacc9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.381182] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Got total of 0 instances {{(pid=62952) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 547.381843] env[62952]: WARNING nova.virt.vmwareapi.driver [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 547.382198] env[62952]: INFO nova.virt.node [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Generated node identity 615127b5-dc82-4491-9c8f-4e6d0caa0690 [ 547.382423] env[62952]: INFO nova.virt.node [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Wrote node identity 615127b5-dc82-4491-9c8f-4e6d0caa0690 to /opt/stack/data/n-cpu-1/compute_id [ 547.885630] env[62952]: WARNING nova.compute.manager [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Compute nodes ['615127b5-dc82-4491-9c8f-4e6d0caa0690'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 548.890861] env[62952]: INFO nova.compute.manager [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 549.896881] env[62952]: WARNING nova.compute.manager [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 549.897250] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.897390] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.897535] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.897686] env[62952]: DEBUG nova.compute.resource_tracker [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 549.898621] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b8d5eb-8dce-4f67-af0d-5a2c79a9a17d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.907020] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b7209d-0e1a-4413-977c-63207980f5a3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.921305] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1c51e0-ab98-477e-b516-504394a7744c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.927715] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c83678d-b106-4a3a-b064-5bbbf3d5fb7d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.956974] env[62952]: DEBUG nova.compute.resource_tracker [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181553MB free_disk=145GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 549.957146] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.957351] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.459807] env[62952]: WARNING nova.compute.resource_tracker [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] No compute node record for cpu-1:615127b5-dc82-4491-9c8f-4e6d0caa0690: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 615127b5-dc82-4491-9c8f-4e6d0caa0690 could not be found. [ 550.964106] env[62952]: INFO nova.compute.resource_tracker [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 615127b5-dc82-4491-9c8f-4e6d0caa0690 [ 552.472603] env[62952]: DEBUG nova.compute.resource_tracker [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 552.472986] env[62952]: DEBUG nova.compute.resource_tracker [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 552.633412] env[62952]: INFO nova.scheduler.client.report [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] [req-cdb43dd0-271d-48b3-9182-bd7cb954ad7e] Created resource provider record via placement API for resource provider with UUID 615127b5-dc82-4491-9c8f-4e6d0caa0690 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 552.649893] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-391c07da-c140-4725-a60a-13b933807a8c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.657268] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28fbae33-802c-4b35-b5c4-832941da7600 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.686551] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a25d80-ff38-4324-b497-88909bcb3ae5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.693824] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e7ed2f9-573f-46be-b623-678cbc03f694 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.706767] env[62952]: DEBUG nova.compute.provider_tree [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 553.244642] env[62952]: DEBUG nova.scheduler.client.report [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Updated inventory for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 553.244877] env[62952]: DEBUG nova.compute.provider_tree [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Updating resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 generation from 0 to 1 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 553.245029] env[62952]: DEBUG nova.compute.provider_tree [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 553.293845] env[62952]: DEBUG nova.compute.provider_tree [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Updating resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 generation from 1 to 2 during operation: update_traits {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 553.798524] env[62952]: DEBUG nova.compute.resource_tracker [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 553.798897] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.841s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.798897] env[62952]: DEBUG nova.service [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Creating RPC server for service compute {{(pid=62952) start /opt/stack/nova/nova/service.py:186}} [ 553.811569] env[62952]: DEBUG nova.service [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] Join ServiceGroup membership for this service compute {{(pid=62952) start /opt/stack/nova/nova/service.py:203}} [ 553.811809] env[62952]: DEBUG nova.servicegroup.drivers.db [None req-2a577da3-5b92-4e0c-bdb0-3fed70052825 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62952) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 589.422670] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "d19a78e1-ac0d-4b3e-9812-d6938298a255" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.422973] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "d19a78e1-ac0d-4b3e-9812-d6938298a255" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.925667] env[62952]: DEBUG nova.compute.manager [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 590.216080] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Acquiring lock "b419be16-cf2e-4ba5-aea3-fd2f56ff006a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.216508] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Lock "b419be16-cf2e-4ba5-aea3-fd2f56ff006a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.472276] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.472563] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.475513] env[62952]: INFO nova.compute.claims [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 590.719519] env[62952]: DEBUG nova.compute.manager [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 591.252644] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.561779] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eddd6b5-a7fc-42a9-9636-64d8ffb22e23 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.572937] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295463e7-38fc-4f79-a6a1-3b0e32446160 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.607421] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa4856e9-534a-4287-ad97-79e5effa25bd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.616150] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3963778-1e5c-481a-aea4-36c4e1b37bf0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.635920] env[62952]: DEBUG nova.compute.provider_tree [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.748615] env[62952]: DEBUG oslo_concurrency.lockutils [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "9c762648-73d9-4976-90e2-13d6a394e7f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.749306] env[62952]: DEBUG oslo_concurrency.lockutils [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "9c762648-73d9-4976-90e2-13d6a394e7f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.140560] env[62952]: DEBUG nova.scheduler.client.report [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.251831] env[62952]: DEBUG nova.compute.manager [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 592.634502] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Acquiring lock "20cad4d8-5a23-47be-8603-2bf576700a14" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.634502] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Lock "20cad4d8-5a23-47be-8603-2bf576700a14" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.647183] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.174s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.648436] env[62952]: DEBUG nova.compute.manager [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 592.652971] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.402s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.654558] env[62952]: INFO nova.compute.claims [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 592.792761] env[62952]: DEBUG oslo_concurrency.lockutils [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.142589] env[62952]: DEBUG nova.compute.manager [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 593.159693] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Acquiring lock "0e49e1b2-076a-4770-bbdc-3d5f7cf67590" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.159914] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Lock "0e49e1b2-076a-4770-bbdc-3d5f7cf67590" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.162352] env[62952]: DEBUG nova.compute.utils [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 593.163969] env[62952]: DEBUG nova.compute.manager [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 593.164211] env[62952]: DEBUG nova.network.neutron [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 593.295606] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Acquiring lock "4c40c507-b988-498c-b68a-a5304541a1c4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.296357] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Lock "4c40c507-b988-498c-b68a-a5304541a1c4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.343595] env[62952]: DEBUG nova.policy [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9f6f0133f4a4ac287d4d8d3f4992754', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1487c06883e444a589d9b688155dc00e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 593.664681] env[62952]: DEBUG nova.compute.manager [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 593.669505] env[62952]: DEBUG nova.compute.manager [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 593.673083] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.779821] env[62952]: DEBUG nova.network.neutron [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Successfully created port: 91d1c070-8a67-4af0-9df8-aa8c642d8552 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 593.798851] env[62952]: DEBUG nova.compute.manager [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 593.807687] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622e42ea-81b5-4f3a-aa90-a1d3fa7b8725 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.818777] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1025b87b-58cf-43b5-bbd2-2be64a40ec5e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.856732] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d3d99a0-b2e3-49c9-9a76-ddaeea69a0bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.866718] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ecf794a-7db1-4b0c-852a-e1f6eaa43406 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.881975] env[62952]: DEBUG nova.compute.provider_tree [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.191458] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.323693] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.389156] env[62952]: DEBUG nova.scheduler.client.report [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.664888] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Acquiring lock "1f51dd77-519f-4d66-9048-5f63e85e3837" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.664888] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Lock "1f51dd77-519f-4d66-9048-5f63e85e3837" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.679931] env[62952]: DEBUG nova.compute.manager [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 594.716986] env[62952]: DEBUG nova.virt.hardware [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 594.717317] env[62952]: DEBUG nova.virt.hardware [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 594.718755] env[62952]: DEBUG nova.virt.hardware [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 594.718755] env[62952]: DEBUG nova.virt.hardware [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 594.718755] env[62952]: DEBUG nova.virt.hardware [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 594.718755] env[62952]: DEBUG nova.virt.hardware [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 594.719745] env[62952]: DEBUG nova.virt.hardware [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 594.719745] env[62952]: DEBUG nova.virt.hardware [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 594.720871] env[62952]: DEBUG nova.virt.hardware [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 594.720871] env[62952]: DEBUG nova.virt.hardware [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 594.720962] env[62952]: DEBUG nova.virt.hardware [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 594.722997] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ac6a720-0c17-46f1-b6dc-59053443668e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.732199] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-503cd4b3-045d-4221-b4d8-fbc6b9e69e83 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.750936] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4834e2-a888-4da4-b3cf-b3cde66a5523 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.897392] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.245s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.898605] env[62952]: DEBUG nova.compute.manager [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 594.901383] env[62952]: DEBUG oslo_concurrency.lockutils [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.110s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.903053] env[62952]: INFO nova.compute.claims [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 595.168077] env[62952]: DEBUG nova.compute.manager [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 595.409542] env[62952]: DEBUG nova.compute.utils [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 595.415383] env[62952]: DEBUG nova.compute.manager [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 595.415383] env[62952]: DEBUG nova.network.neutron [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 595.462194] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Acquiring lock "5db9f2cd-faf1-4972-8fa0-7b7705f40bf4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.462447] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Lock "5db9f2cd-faf1-4972-8fa0-7b7705f40bf4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.651527] env[62952]: DEBUG nova.policy [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aa5d996a798b4430a9f6cb57acf987fb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f83de16a83f14816a0085bcb8d110a11', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 595.705257] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.915469] env[62952]: DEBUG nova.compute.manager [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 595.966784] env[62952]: DEBUG nova.compute.manager [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 596.120330] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcf5993f-6d2d-48e7-ae79-8e0660933027 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.128290] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9b36077-52fc-4ed4-ae0b-d600011c092f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.160801] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3553a062-ce95-4a83-a6de-08d8b18c2980 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.169137] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9edeac53-dccd-4e6d-8bdf-6649669eaa04 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.186606] env[62952]: DEBUG nova.compute.provider_tree [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.492964] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.606798] env[62952]: ERROR nova.compute.manager [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 91d1c070-8a67-4af0-9df8-aa8c642d8552, please check neutron logs for more information. [ 596.606798] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 596.606798] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.606798] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 596.606798] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 596.606798] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 596.606798] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 596.606798] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 596.606798] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.606798] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 596.606798] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.606798] env[62952]: ERROR nova.compute.manager raise self.value [ 596.606798] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 596.606798] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 596.606798] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.606798] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 596.607388] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.607388] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 596.607388] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 91d1c070-8a67-4af0-9df8-aa8c642d8552, please check neutron logs for more information. [ 596.607388] env[62952]: ERROR nova.compute.manager [ 596.609846] env[62952]: Traceback (most recent call last): [ 596.609948] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 596.609948] env[62952]: listener.cb(fileno) [ 596.609948] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.609948] env[62952]: result = function(*args, **kwargs) [ 596.609948] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 596.609948] env[62952]: return func(*args, **kwargs) [ 596.609948] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.609948] env[62952]: raise e [ 596.609948] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.609948] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 596.609948] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 596.609948] env[62952]: created_port_ids = self._update_ports_for_instance( [ 596.609948] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 596.609948] env[62952]: with excutils.save_and_reraise_exception(): [ 596.609948] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.609948] env[62952]: self.force_reraise() [ 596.609948] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.609948] env[62952]: raise self.value [ 596.609948] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 596.609948] env[62952]: updated_port = self._update_port( [ 596.609948] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.609948] env[62952]: _ensure_no_port_binding_failure(port) [ 596.609948] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.609948] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 596.610872] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 91d1c070-8a67-4af0-9df8-aa8c642d8552, please check neutron logs for more information. [ 596.610872] env[62952]: Removing descriptor: 15 [ 596.613830] env[62952]: ERROR nova.compute.manager [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 91d1c070-8a67-4af0-9df8-aa8c642d8552, please check neutron logs for more information. [ 596.613830] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Traceback (most recent call last): [ 596.613830] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 596.613830] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] yield resources [ 596.613830] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 596.613830] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] self.driver.spawn(context, instance, image_meta, [ 596.613830] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 596.613830] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.613830] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.613830] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] vm_ref = self.build_virtual_machine(instance, [ 596.613830] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.614367] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.614367] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.614367] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] for vif in network_info: [ 596.614367] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 596.614367] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] return self._sync_wrapper(fn, *args, **kwargs) [ 596.614367] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 596.614367] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] self.wait() [ 596.614367] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 596.614367] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] self[:] = self._gt.wait() [ 596.614367] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.614367] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] return self._exit_event.wait() [ 596.614367] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.614367] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] result = hub.switch() [ 596.614742] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.614742] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] return self.greenlet.switch() [ 596.614742] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.614742] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] result = function(*args, **kwargs) [ 596.614742] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 596.614742] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] return func(*args, **kwargs) [ 596.614742] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.614742] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] raise e [ 596.614742] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.614742] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] nwinfo = self.network_api.allocate_for_instance( [ 596.614742] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 596.614742] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] created_port_ids = self._update_ports_for_instance( [ 596.614742] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 596.615694] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] with excutils.save_and_reraise_exception(): [ 596.615694] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.615694] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] self.force_reraise() [ 596.615694] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.615694] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] raise self.value [ 596.615694] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 596.615694] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] updated_port = self._update_port( [ 596.615694] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.615694] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] _ensure_no_port_binding_failure(port) [ 596.615694] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.615694] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] raise exception.PortBindingFailed(port_id=port['id']) [ 596.615694] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] nova.exception.PortBindingFailed: Binding failed for port 91d1c070-8a67-4af0-9df8-aa8c642d8552, please check neutron logs for more information. [ 596.615694] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] [ 596.616058] env[62952]: INFO nova.compute.manager [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Terminating instance [ 596.620178] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "refresh_cache-d19a78e1-ac0d-4b3e-9812-d6938298a255" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.620178] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquired lock "refresh_cache-d19a78e1-ac0d-4b3e-9812-d6938298a255" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.620178] env[62952]: DEBUG nova.network.neutron [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 596.692138] env[62952]: DEBUG nova.scheduler.client.report [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 596.928273] env[62952]: DEBUG nova.compute.manager [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 596.962841] env[62952]: DEBUG nova.virt.hardware [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 596.962985] env[62952]: DEBUG nova.virt.hardware [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 596.963217] env[62952]: DEBUG nova.virt.hardware [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 596.963292] env[62952]: DEBUG nova.virt.hardware [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 596.963425] env[62952]: DEBUG nova.virt.hardware [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 596.963616] env[62952]: DEBUG nova.virt.hardware [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 596.964142] env[62952]: DEBUG nova.virt.hardware [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 596.964142] env[62952]: DEBUG nova.virt.hardware [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 596.964387] env[62952]: DEBUG nova.virt.hardware [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 596.964737] env[62952]: DEBUG nova.virt.hardware [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 596.965209] env[62952]: DEBUG nova.virt.hardware [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 596.966054] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ee87c33-cc19-49f1-9e4e-45ee002033f5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.975859] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19abcba5-fb75-4a27-a09a-7f719bc397ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.990372] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Acquiring lock "dea74836-bdbb-4260-b15d-2dab04ee3940" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.990599] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Lock "dea74836-bdbb-4260-b15d-2dab04ee3940" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.185483] env[62952]: DEBUG nova.network.neutron [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.201750] env[62952]: DEBUG oslo_concurrency.lockutils [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.300s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.201870] env[62952]: DEBUG nova.compute.manager [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 597.205457] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.532s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.210172] env[62952]: INFO nova.compute.claims [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 597.364390] env[62952]: DEBUG nova.compute.manager [req-bffffd52-4fdf-4dbe-90b8-ae2e27cb6641 req-0f35b18e-81b1-4e76-a9f9-c0f7fc8d6547 service nova] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Received event network-changed-91d1c070-8a67-4af0-9df8-aa8c642d8552 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 597.364576] env[62952]: DEBUG nova.compute.manager [req-bffffd52-4fdf-4dbe-90b8-ae2e27cb6641 req-0f35b18e-81b1-4e76-a9f9-c0f7fc8d6547 service nova] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Refreshing instance network info cache due to event network-changed-91d1c070-8a67-4af0-9df8-aa8c642d8552. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 597.364829] env[62952]: DEBUG oslo_concurrency.lockutils [req-bffffd52-4fdf-4dbe-90b8-ae2e27cb6641 req-0f35b18e-81b1-4e76-a9f9-c0f7fc8d6547 service nova] Acquiring lock "refresh_cache-d19a78e1-ac0d-4b3e-9812-d6938298a255" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.502633] env[62952]: DEBUG nova.compute.manager [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 597.535985] env[62952]: DEBUG nova.network.neutron [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.717477] env[62952]: DEBUG nova.compute.utils [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 597.721275] env[62952]: DEBUG nova.compute.manager [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 597.721275] env[62952]: DEBUG nova.network.neutron [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 597.784617] env[62952]: DEBUG nova.policy [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6967f8ebdc944dbfb2d19b5d7acdd3e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91fde3ceb94649f698459dfc174bf835', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 597.906189] env[62952]: DEBUG nova.network.neutron [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Successfully created port: 6a9a4e0e-736f-4156-8381-eb26aaf6028c {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 598.039264] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.039772] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Releasing lock "refresh_cache-d19a78e1-ac0d-4b3e-9812-d6938298a255" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.040216] env[62952]: DEBUG nova.compute.manager [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 598.040403] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 598.040692] env[62952]: DEBUG oslo_concurrency.lockutils [req-bffffd52-4fdf-4dbe-90b8-ae2e27cb6641 req-0f35b18e-81b1-4e76-a9f9-c0f7fc8d6547 service nova] Acquired lock "refresh_cache-d19a78e1-ac0d-4b3e-9812-d6938298a255" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.040914] env[62952]: DEBUG nova.network.neutron [req-bffffd52-4fdf-4dbe-90b8-ae2e27cb6641 req-0f35b18e-81b1-4e76-a9f9-c0f7fc8d6547 service nova] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Refreshing network info cache for port 91d1c070-8a67-4af0-9df8-aa8c642d8552 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 598.042761] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e732541f-cd21-424e-9f3c-a079c376da78 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.057216] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf523ad9-ffa0-4b0a-a2a3-8533e8e9ec19 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.084620] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d19a78e1-ac0d-4b3e-9812-d6938298a255 could not be found. [ 598.084897] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 598.085266] env[62952]: INFO nova.compute.manager [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Took 0.04 seconds to destroy the instance on the hypervisor. [ 598.085633] env[62952]: DEBUG oslo.service.loopingcall [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 598.085804] env[62952]: DEBUG nova.compute.manager [-] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.085902] env[62952]: DEBUG nova.network.neutron [-] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 598.147740] env[62952]: DEBUG nova.network.neutron [-] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.225856] env[62952]: DEBUG nova.compute.manager [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 598.436669] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03bc1f05-5c7a-4720-b31a-b8b3604dfaca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.444353] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9570a2-1638-4790-9e19-9268a879e804 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.478124] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85309075-a5ab-4dfa-9970-f9cb8ef3c84e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.485749] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9639ad89-0ef2-4a4f-8c0c-1b94502a407d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.499964] env[62952]: DEBUG nova.compute.provider_tree [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.600239] env[62952]: DEBUG nova.network.neutron [req-bffffd52-4fdf-4dbe-90b8-ae2e27cb6641 req-0f35b18e-81b1-4e76-a9f9-c0f7fc8d6547 service nova] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.652824] env[62952]: DEBUG nova.network.neutron [-] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.812219] env[62952]: DEBUG nova.network.neutron [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Successfully created port: 5f1fb2a9-4d00-45a4-b6f3-6a9766d92324 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 598.938728] env[62952]: DEBUG nova.network.neutron [req-bffffd52-4fdf-4dbe-90b8-ae2e27cb6641 req-0f35b18e-81b1-4e76-a9f9-c0f7fc8d6547 service nova] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.003338] env[62952]: DEBUG nova.scheduler.client.report [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.158029] env[62952]: INFO nova.compute.manager [-] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Took 1.07 seconds to deallocate network for instance. [ 599.162184] env[62952]: DEBUG nova.compute.claims [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 599.162380] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.245833] env[62952]: DEBUG nova.compute.manager [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 599.285889] env[62952]: DEBUG nova.virt.hardware [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 599.286195] env[62952]: DEBUG nova.virt.hardware [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 599.286348] env[62952]: DEBUG nova.virt.hardware [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 599.286524] env[62952]: DEBUG nova.virt.hardware [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 599.286969] env[62952]: DEBUG nova.virt.hardware [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 599.287196] env[62952]: DEBUG nova.virt.hardware [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 599.287429] env[62952]: DEBUG nova.virt.hardware [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 599.287585] env[62952]: DEBUG nova.virt.hardware [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 599.287742] env[62952]: DEBUG nova.virt.hardware [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 599.287895] env[62952]: DEBUG nova.virt.hardware [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 599.288127] env[62952]: DEBUG nova.virt.hardware [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 599.289533] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdba62c2-146a-4f1b-b87c-4d0801366c4a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.298954] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b92003bd-baf5-477d-b99f-8c411ee17bc9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.441321] env[62952]: DEBUG oslo_concurrency.lockutils [req-bffffd52-4fdf-4dbe-90b8-ae2e27cb6641 req-0f35b18e-81b1-4e76-a9f9-c0f7fc8d6547 service nova] Releasing lock "refresh_cache-d19a78e1-ac0d-4b3e-9812-d6938298a255" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.510509] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.304s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.510925] env[62952]: DEBUG nova.compute.manager [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 599.514059] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.323s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.515810] env[62952]: INFO nova.compute.claims [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.870836] env[62952]: DEBUG nova.compute.manager [req-09c9fe9b-f702-4c18-9361-a5ea0dea4dc5 req-bc3000c3-efc3-4c7f-ba0c-01c6134b174e service nova] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Received event network-vif-deleted-91d1c070-8a67-4af0-9df8-aa8c642d8552 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 600.020680] env[62952]: DEBUG nova.compute.utils [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 600.025066] env[62952]: DEBUG nova.compute.manager [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 600.025066] env[62952]: DEBUG nova.network.neutron [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 600.447611] env[62952]: DEBUG nova.policy [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9629d9a1d29644d4b4b6f3bdf071974a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd357aa1aa3a34557ba711ca26fe1ad39', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 600.525581] env[62952]: DEBUG nova.compute.manager [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 600.712903] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d65c74dc-97ff-4c75-9bb9-166fbf0b274e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.721173] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992e8263-8014-4a77-b3ec-d3ed3f200919 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.769484] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be75ef7e-51f5-4962-b830-62cef9691f74 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.779281] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5b1345c-1cb0-49a5-b84b-d9c05f1250bd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.793413] env[62952]: DEBUG nova.compute.provider_tree [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.298115] env[62952]: DEBUG nova.scheduler.client.report [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.542069] env[62952]: DEBUG nova.compute.manager [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 601.586811] env[62952]: DEBUG nova.virt.hardware [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 601.586811] env[62952]: DEBUG nova.virt.hardware [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 601.587360] env[62952]: DEBUG nova.virt.hardware [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 601.587360] env[62952]: DEBUG nova.virt.hardware [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 601.587795] env[62952]: DEBUG nova.virt.hardware [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 601.587795] env[62952]: DEBUG nova.virt.hardware [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 601.588026] env[62952]: DEBUG nova.virt.hardware [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 601.588156] env[62952]: DEBUG nova.virt.hardware [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 601.588384] env[62952]: DEBUG nova.virt.hardware [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 601.588603] env[62952]: DEBUG nova.virt.hardware [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 601.588839] env[62952]: DEBUG nova.virt.hardware [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 601.590054] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7045e0c0-40b5-4786-93b5-242dbe902491 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.602880] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee86f469-cdb0-4a2a-acfa-c8a7d7ecd865 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.810322] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.296s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.811204] env[62952]: DEBUG nova.compute.manager [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 601.815503] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.492s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.818084] env[62952]: INFO nova.compute.claims [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 602.094663] env[62952]: DEBUG nova.network.neutron [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Successfully created port: f2e7e01f-0ace-4b09-a54a-cbc7a33c5f3f {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 602.317595] env[62952]: DEBUG nova.compute.utils [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 602.319692] env[62952]: DEBUG nova.compute.manager [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 602.319692] env[62952]: DEBUG nova.network.neutron [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 602.590134] env[62952]: DEBUG nova.policy [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0cf74331d0d64e489d9a935e07f34c17', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '01ccf3d5c23f414bae289a71a502f4d1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 602.813673] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 602.814155] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 602.814884] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 602.814884] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 602.823153] env[62952]: DEBUG nova.compute.manager [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 602.860888] env[62952]: ERROR nova.compute.manager [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6a9a4e0e-736f-4156-8381-eb26aaf6028c, please check neutron logs for more information. [ 602.860888] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 602.860888] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.860888] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 602.860888] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.860888] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 602.860888] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.860888] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 602.860888] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.860888] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 602.860888] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.860888] env[62952]: ERROR nova.compute.manager raise self.value [ 602.860888] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.860888] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 602.860888] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.860888] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 602.861443] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.861443] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 602.861443] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6a9a4e0e-736f-4156-8381-eb26aaf6028c, please check neutron logs for more information. [ 602.861443] env[62952]: ERROR nova.compute.manager [ 602.861443] env[62952]: Traceback (most recent call last): [ 602.861583] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 602.861583] env[62952]: listener.cb(fileno) [ 602.861583] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.861583] env[62952]: result = function(*args, **kwargs) [ 602.861583] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.861583] env[62952]: return func(*args, **kwargs) [ 602.861583] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.861583] env[62952]: raise e [ 602.861583] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.861583] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 602.861583] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.861583] env[62952]: created_port_ids = self._update_ports_for_instance( [ 602.861583] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.861583] env[62952]: with excutils.save_and_reraise_exception(): [ 602.861583] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.861583] env[62952]: self.force_reraise() [ 602.861583] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.861583] env[62952]: raise self.value [ 602.861583] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.861583] env[62952]: updated_port = self._update_port( [ 602.861583] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.861583] env[62952]: _ensure_no_port_binding_failure(port) [ 602.861583] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.861583] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 602.861583] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 6a9a4e0e-736f-4156-8381-eb26aaf6028c, please check neutron logs for more information. [ 602.861583] env[62952]: Removing descriptor: 16 [ 602.866111] env[62952]: ERROR nova.compute.manager [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6a9a4e0e-736f-4156-8381-eb26aaf6028c, please check neutron logs for more information. [ 602.866111] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Traceback (most recent call last): [ 602.866111] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 602.866111] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] yield resources [ 602.866111] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.866111] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] self.driver.spawn(context, instance, image_meta, [ 602.866111] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 602.866111] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.866111] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.866111] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] vm_ref = self.build_virtual_machine(instance, [ 602.866111] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.866467] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.866467] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.866467] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] for vif in network_info: [ 602.866467] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.866467] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] return self._sync_wrapper(fn, *args, **kwargs) [ 602.866467] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.866467] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] self.wait() [ 602.866467] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.866467] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] self[:] = self._gt.wait() [ 602.866467] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.866467] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] return self._exit_event.wait() [ 602.866467] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.866467] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] result = hub.switch() [ 602.866858] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.866858] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] return self.greenlet.switch() [ 602.866858] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.866858] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] result = function(*args, **kwargs) [ 602.866858] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.866858] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] return func(*args, **kwargs) [ 602.866858] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.866858] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] raise e [ 602.866858] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.866858] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] nwinfo = self.network_api.allocate_for_instance( [ 602.866858] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.866858] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] created_port_ids = self._update_ports_for_instance( [ 602.866858] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.867233] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] with excutils.save_and_reraise_exception(): [ 602.867233] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.867233] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] self.force_reraise() [ 602.867233] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.867233] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] raise self.value [ 602.867233] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.867233] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] updated_port = self._update_port( [ 602.867233] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.867233] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] _ensure_no_port_binding_failure(port) [ 602.867233] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.867233] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] raise exception.PortBindingFailed(port_id=port['id']) [ 602.867233] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] nova.exception.PortBindingFailed: Binding failed for port 6a9a4e0e-736f-4156-8381-eb26aaf6028c, please check neutron logs for more information. [ 602.867233] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] [ 602.867558] env[62952]: INFO nova.compute.manager [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Terminating instance [ 602.867558] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Acquiring lock "refresh_cache-b419be16-cf2e-4ba5-aea3-fd2f56ff006a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.867558] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Acquired lock "refresh_cache-b419be16-cf2e-4ba5-aea3-fd2f56ff006a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.867558] env[62952]: DEBUG nova.network.neutron [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 603.153644] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0386aee9-76ec-4b5f-b328-60e5281dc345 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.161978] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1700ebb-f768-4710-94b2-0270d7714e60 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.200767] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-770b7498-8fbf-4cf3-a1f8-cbf214aecd55 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.208729] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e595543-077b-4ed5-83ef-f6b0f1073d06 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.222735] env[62952]: DEBUG nova.compute.provider_tree [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.320398] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 603.320398] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 603.320398] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 603.320398] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 603.320398] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 603.320398] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 603.320637] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Didn't find any instances for network info cache update. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 603.320637] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.327298] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.327298] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.327298] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.327298] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.327298] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.600833] env[62952]: DEBUG nova.network.neutron [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.665951] env[62952]: DEBUG nova.compute.manager [req-1a5963ec-c95f-4446-9f5e-7f8e52cf36ee req-3234e381-9598-4654-9ecb-7fa9cf8cc8d8 service nova] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Received event network-changed-6a9a4e0e-736f-4156-8381-eb26aaf6028c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 603.665951] env[62952]: DEBUG nova.compute.manager [req-1a5963ec-c95f-4446-9f5e-7f8e52cf36ee req-3234e381-9598-4654-9ecb-7fa9cf8cc8d8 service nova] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Refreshing instance network info cache due to event network-changed-6a9a4e0e-736f-4156-8381-eb26aaf6028c. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 603.665951] env[62952]: DEBUG oslo_concurrency.lockutils [req-1a5963ec-c95f-4446-9f5e-7f8e52cf36ee req-3234e381-9598-4654-9ecb-7fa9cf8cc8d8 service nova] Acquiring lock "refresh_cache-b419be16-cf2e-4ba5-aea3-fd2f56ff006a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.728027] env[62952]: DEBUG nova.scheduler.client.report [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.834829] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Getting list of instances from cluster (obj){ [ 603.834829] env[62952]: value = "domain-c8" [ 603.834829] env[62952]: _type = "ClusterComputeResource" [ 603.834829] env[62952]: } {{(pid=62952) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 603.836410] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2804ec35-66dd-4b29-b936-40c07096791b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.844816] env[62952]: DEBUG nova.network.neutron [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.849484] env[62952]: DEBUG nova.compute.manager [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 603.861042] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Got total of 0 instances {{(pid=62952) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 603.861042] env[62952]: WARNING nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] While synchronizing instance power states, found 6 instances in the database and 0 instances on the hypervisor. [ 603.861042] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Triggering sync for uuid d19a78e1-ac0d-4b3e-9812-d6938298a255 {{(pid=62952) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 603.869382] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Triggering sync for uuid b419be16-cf2e-4ba5-aea3-fd2f56ff006a {{(pid=62952) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 603.869707] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Triggering sync for uuid 9c762648-73d9-4976-90e2-13d6a394e7f6 {{(pid=62952) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 603.869945] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Triggering sync for uuid 20cad4d8-5a23-47be-8603-2bf576700a14 {{(pid=62952) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 603.870089] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Triggering sync for uuid 0e49e1b2-076a-4770-bbdc-3d5f7cf67590 {{(pid=62952) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 603.870287] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Triggering sync for uuid 4c40c507-b988-498c-b68a-a5304541a1c4 {{(pid=62952) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 603.871255] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "d19a78e1-ac0d-4b3e-9812-d6938298a255" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.872580] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "b419be16-cf2e-4ba5-aea3-fd2f56ff006a" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.873319] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "9c762648-73d9-4976-90e2-13d6a394e7f6" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.873319] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "20cad4d8-5a23-47be-8603-2bf576700a14" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.873319] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "0e49e1b2-076a-4770-bbdc-3d5f7cf67590" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.873456] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "4c40c507-b988-498c-b68a-a5304541a1c4" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.876760] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.877050] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 603.878864] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.909297] env[62952]: DEBUG nova.virt.hardware [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 603.909297] env[62952]: DEBUG nova.virt.hardware [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 603.909297] env[62952]: DEBUG nova.virt.hardware [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 603.909480] env[62952]: DEBUG nova.virt.hardware [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 603.909561] env[62952]: DEBUG nova.virt.hardware [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 603.909887] env[62952]: DEBUG nova.virt.hardware [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 603.909988] env[62952]: DEBUG nova.virt.hardware [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 603.915570] env[62952]: DEBUG nova.virt.hardware [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 603.916590] env[62952]: DEBUG nova.virt.hardware [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 603.916590] env[62952]: DEBUG nova.virt.hardware [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 603.916590] env[62952]: DEBUG nova.virt.hardware [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 603.922869] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ca3f3cd-a1fe-4d6f-af45-20db3131e546 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.928196] env[62952]: DEBUG nova.compute.manager [req-10514416-feda-4286-ad8d-f071e69aa100 req-073165ca-b271-4466-a8fe-3980b226a40b service nova] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Received event network-changed-5f1fb2a9-4d00-45a4-b6f3-6a9766d92324 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 603.928196] env[62952]: DEBUG nova.compute.manager [req-10514416-feda-4286-ad8d-f071e69aa100 req-073165ca-b271-4466-a8fe-3980b226a40b service nova] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Refreshing instance network info cache due to event network-changed-5f1fb2a9-4d00-45a4-b6f3-6a9766d92324. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 603.928196] env[62952]: DEBUG oslo_concurrency.lockutils [req-10514416-feda-4286-ad8d-f071e69aa100 req-073165ca-b271-4466-a8fe-3980b226a40b service nova] Acquiring lock "refresh_cache-9c762648-73d9-4976-90e2-13d6a394e7f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.929719] env[62952]: DEBUG oslo_concurrency.lockutils [req-10514416-feda-4286-ad8d-f071e69aa100 req-073165ca-b271-4466-a8fe-3980b226a40b service nova] Acquired lock "refresh_cache-9c762648-73d9-4976-90e2-13d6a394e7f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.929719] env[62952]: DEBUG nova.network.neutron [req-10514416-feda-4286-ad8d-f071e69aa100 req-073165ca-b271-4466-a8fe-3980b226a40b service nova] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Refreshing network info cache for port 5f1fb2a9-4d00-45a4-b6f3-6a9766d92324 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 603.943464] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a94f9fa6-26ae-4e1d-9825-2d7ac035f973 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.205646] env[62952]: DEBUG nova.network.neutron [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Successfully created port: 0c947d5e-7083-46d4-9b46-6e4a7331b432 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 604.230028] env[62952]: ERROR nova.compute.manager [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5f1fb2a9-4d00-45a4-b6f3-6a9766d92324, please check neutron logs for more information. [ 604.230028] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 604.230028] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.230028] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 604.230028] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 604.230028] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 604.230028] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 604.230028] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 604.230028] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.230028] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 604.230028] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.230028] env[62952]: ERROR nova.compute.manager raise self.value [ 604.230028] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 604.230028] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 604.230028] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.230028] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 604.230635] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.230635] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 604.230635] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5f1fb2a9-4d00-45a4-b6f3-6a9766d92324, please check neutron logs for more information. [ 604.230635] env[62952]: ERROR nova.compute.manager [ 604.230635] env[62952]: Traceback (most recent call last): [ 604.230635] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 604.230635] env[62952]: listener.cb(fileno) [ 604.230635] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.230635] env[62952]: result = function(*args, **kwargs) [ 604.230635] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.230635] env[62952]: return func(*args, **kwargs) [ 604.230635] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.230635] env[62952]: raise e [ 604.230635] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.230635] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 604.230635] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 604.230635] env[62952]: created_port_ids = self._update_ports_for_instance( [ 604.230635] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 604.230635] env[62952]: with excutils.save_and_reraise_exception(): [ 604.230635] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.230635] env[62952]: self.force_reraise() [ 604.230635] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.230635] env[62952]: raise self.value [ 604.230635] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 604.230635] env[62952]: updated_port = self._update_port( [ 604.230635] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.230635] env[62952]: _ensure_no_port_binding_failure(port) [ 604.230635] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.230635] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 604.231418] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 5f1fb2a9-4d00-45a4-b6f3-6a9766d92324, please check neutron logs for more information. [ 604.231418] env[62952]: Removing descriptor: 15 [ 604.231418] env[62952]: ERROR nova.compute.manager [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5f1fb2a9-4d00-45a4-b6f3-6a9766d92324, please check neutron logs for more information. [ 604.231418] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Traceback (most recent call last): [ 604.231418] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 604.231418] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] yield resources [ 604.231418] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.231418] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] self.driver.spawn(context, instance, image_meta, [ 604.231418] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 604.231418] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.231418] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.231418] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] vm_ref = self.build_virtual_machine(instance, [ 604.231768] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.231768] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.231768] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.231768] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] for vif in network_info: [ 604.231768] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.231768] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] return self._sync_wrapper(fn, *args, **kwargs) [ 604.231768] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.231768] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] self.wait() [ 604.231768] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.231768] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] self[:] = self._gt.wait() [ 604.231768] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.231768] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] return self._exit_event.wait() [ 604.231768] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.232159] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] result = hub.switch() [ 604.232159] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.232159] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] return self.greenlet.switch() [ 604.232159] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.232159] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] result = function(*args, **kwargs) [ 604.232159] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.232159] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] return func(*args, **kwargs) [ 604.232159] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.232159] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] raise e [ 604.232159] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.232159] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] nwinfo = self.network_api.allocate_for_instance( [ 604.232159] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 604.232159] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] created_port_ids = self._update_ports_for_instance( [ 604.232691] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 604.232691] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] with excutils.save_and_reraise_exception(): [ 604.232691] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.232691] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] self.force_reraise() [ 604.232691] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.232691] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] raise self.value [ 604.232691] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 604.232691] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] updated_port = self._update_port( [ 604.232691] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.232691] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] _ensure_no_port_binding_failure(port) [ 604.232691] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.232691] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] raise exception.PortBindingFailed(port_id=port['id']) [ 604.233046] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] nova.exception.PortBindingFailed: Binding failed for port 5f1fb2a9-4d00-45a4-b6f3-6a9766d92324, please check neutron logs for more information. [ 604.233046] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] [ 604.233046] env[62952]: INFO nova.compute.manager [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Terminating instance [ 604.237019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "refresh_cache-9c762648-73d9-4976-90e2-13d6a394e7f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.237019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.237019] env[62952]: DEBUG nova.compute.manager [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 604.238626] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.534s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.240342] env[62952]: INFO nova.compute.claims [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.354148] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Releasing lock "refresh_cache-b419be16-cf2e-4ba5-aea3-fd2f56ff006a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.357187] env[62952]: DEBUG nova.compute.manager [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 604.357187] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 604.357187] env[62952]: DEBUG oslo_concurrency.lockutils [req-1a5963ec-c95f-4446-9f5e-7f8e52cf36ee req-3234e381-9598-4654-9ecb-7fa9cf8cc8d8 service nova] Acquired lock "refresh_cache-b419be16-cf2e-4ba5-aea3-fd2f56ff006a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.357187] env[62952]: DEBUG nova.network.neutron [req-1a5963ec-c95f-4446-9f5e-7f8e52cf36ee req-3234e381-9598-4654-9ecb-7fa9cf8cc8d8 service nova] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Refreshing network info cache for port 6a9a4e0e-736f-4156-8381-eb26aaf6028c {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 604.357187] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a74d2c51-64e4-4936-89ee-0255f4372bec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.374268] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13c882c9-ae42-4025-9af1-67291455915f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.391982] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.402223] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b419be16-cf2e-4ba5-aea3-fd2f56ff006a could not be found. [ 604.402530] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 604.403235] env[62952]: INFO nova.compute.manager [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Took 0.05 seconds to destroy the instance on the hypervisor. [ 604.403785] env[62952]: DEBUG oslo.service.loopingcall [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 604.404076] env[62952]: DEBUG nova.compute.manager [-] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 604.404198] env[62952]: DEBUG nova.network.neutron [-] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 604.471286] env[62952]: DEBUG nova.network.neutron [-] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.518289] env[62952]: DEBUG nova.network.neutron [req-10514416-feda-4286-ad8d-f071e69aa100 req-073165ca-b271-4466-a8fe-3980b226a40b service nova] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.743052] env[62952]: DEBUG nova.compute.utils [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 604.743052] env[62952]: DEBUG nova.compute.manager [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 604.743052] env[62952]: DEBUG nova.network.neutron [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 604.912256] env[62952]: DEBUG nova.network.neutron [req-1a5963ec-c95f-4446-9f5e-7f8e52cf36ee req-3234e381-9598-4654-9ecb-7fa9cf8cc8d8 service nova] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.976230] env[62952]: DEBUG nova.network.neutron [-] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.996476] env[62952]: DEBUG nova.network.neutron [req-10514416-feda-4286-ad8d-f071e69aa100 req-073165ca-b271-4466-a8fe-3980b226a40b service nova] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.004173] env[62952]: DEBUG nova.policy [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4791c5fe6ec422dac73219d15153ceb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '05d75f70c9d64cd8b0a390cb8a6ee929', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 605.208078] env[62952]: DEBUG nova.network.neutron [req-1a5963ec-c95f-4446-9f5e-7f8e52cf36ee req-3234e381-9598-4654-9ecb-7fa9cf8cc8d8 service nova] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.247980] env[62952]: DEBUG nova.compute.manager [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 605.432722] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec4bb0ac-d218-479a-998a-f8e1e09ab315 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.442029] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f538c5-f79a-470d-a4f3-37c866503fe2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.475733] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2535c8c-eff8-4da0-81bf-439c745dc14d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.483868] env[62952]: INFO nova.compute.manager [-] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Took 1.08 seconds to deallocate network for instance. [ 605.487461] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c5b8a4-d4e6-47ca-b2c0-c08fd68f4b86 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.492942] env[62952]: DEBUG nova.compute.claims [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 605.493535] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.503824] env[62952]: DEBUG oslo_concurrency.lockutils [req-10514416-feda-4286-ad8d-f071e69aa100 req-073165ca-b271-4466-a8fe-3980b226a40b service nova] Releasing lock "refresh_cache-9c762648-73d9-4976-90e2-13d6a394e7f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.504419] env[62952]: DEBUG nova.compute.provider_tree [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.506301] env[62952]: DEBUG oslo_concurrency.lockutils [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "refresh_cache-9c762648-73d9-4976-90e2-13d6a394e7f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.506770] env[62952]: DEBUG nova.network.neutron [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 605.711244] env[62952]: DEBUG oslo_concurrency.lockutils [req-1a5963ec-c95f-4446-9f5e-7f8e52cf36ee req-3234e381-9598-4654-9ecb-7fa9cf8cc8d8 service nova] Releasing lock "refresh_cache-b419be16-cf2e-4ba5-aea3-fd2f56ff006a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.014415] env[62952]: DEBUG nova.scheduler.client.report [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.066730] env[62952]: DEBUG nova.network.neutron [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.141347] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Acquiring lock "41dbbdcf-667c-4969-88da-43e42840ed20" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.142778] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Lock "41dbbdcf-667c-4969-88da-43e42840ed20" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.250859] env[62952]: DEBUG nova.network.neutron [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Successfully created port: d5a318e4-4330-4101-a6d4-803e4a6f63ae {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 606.260202] env[62952]: DEBUG nova.compute.manager [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 606.284205] env[62952]: DEBUG nova.virt.hardware [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 606.284463] env[62952]: DEBUG nova.virt.hardware [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 606.284619] env[62952]: DEBUG nova.virt.hardware [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 606.284858] env[62952]: DEBUG nova.virt.hardware [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 606.285070] env[62952]: DEBUG nova.virt.hardware [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 606.285228] env[62952]: DEBUG nova.virt.hardware [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 606.285437] env[62952]: DEBUG nova.virt.hardware [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 606.285630] env[62952]: DEBUG nova.virt.hardware [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 606.286109] env[62952]: DEBUG nova.virt.hardware [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 606.286109] env[62952]: DEBUG nova.virt.hardware [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 606.286218] env[62952]: DEBUG nova.virt.hardware [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 606.287137] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fed25f1-d198-473f-b522-3d36c6dbbd68 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.297286] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f87947ac-58e7-4487-a457-a16e94aa08eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.391409] env[62952]: DEBUG nova.network.neutron [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.525404] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.287s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.525922] env[62952]: DEBUG nova.compute.manager [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 606.529030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.036s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.530836] env[62952]: INFO nova.compute.claims [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.644407] env[62952]: DEBUG nova.compute.manager [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 606.894447] env[62952]: DEBUG oslo_concurrency.lockutils [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "refresh_cache-9c762648-73d9-4976-90e2-13d6a394e7f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.895302] env[62952]: DEBUG nova.compute.manager [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 606.895518] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 606.895827] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-815bfa10-832d-4afe-8a87-bfb7c98da980 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.905945] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-172e2fd7-90a7-454b-bdcc-a4d6c55e406a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.931154] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9c762648-73d9-4976-90e2-13d6a394e7f6 could not be found. [ 606.931930] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 606.932069] env[62952]: INFO nova.compute.manager [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 606.932360] env[62952]: DEBUG oslo.service.loopingcall [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 606.932821] env[62952]: DEBUG nova.compute.manager [-] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.932950] env[62952]: DEBUG nova.network.neutron [-] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 606.988382] env[62952]: DEBUG nova.network.neutron [-] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.035178] env[62952]: DEBUG nova.compute.utils [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 607.040021] env[62952]: DEBUG nova.compute.manager [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 607.040021] env[62952]: DEBUG nova.network.neutron [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 607.185080] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.188768] env[62952]: DEBUG nova.policy [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de4b33ebcbcf459eba2953ef70c66f1e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f47257114d1548db8ae9b9b2da345a7d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 607.338384] env[62952]: ERROR nova.compute.manager [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f2e7e01f-0ace-4b09-a54a-cbc7a33c5f3f, please check neutron logs for more information. [ 607.338384] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 607.338384] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.338384] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 607.338384] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.338384] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 607.338384] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.338384] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 607.338384] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.338384] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 607.338384] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.338384] env[62952]: ERROR nova.compute.manager raise self.value [ 607.338384] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.338384] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 607.338384] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.338384] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 607.338915] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.338915] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 607.338915] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f2e7e01f-0ace-4b09-a54a-cbc7a33c5f3f, please check neutron logs for more information. [ 607.338915] env[62952]: ERROR nova.compute.manager [ 607.338915] env[62952]: Traceback (most recent call last): [ 607.338915] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 607.338915] env[62952]: listener.cb(fileno) [ 607.338915] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.338915] env[62952]: result = function(*args, **kwargs) [ 607.338915] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.338915] env[62952]: return func(*args, **kwargs) [ 607.338915] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.338915] env[62952]: raise e [ 607.338915] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.338915] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 607.338915] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.338915] env[62952]: created_port_ids = self._update_ports_for_instance( [ 607.338915] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.338915] env[62952]: with excutils.save_and_reraise_exception(): [ 607.338915] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.338915] env[62952]: self.force_reraise() [ 607.338915] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.338915] env[62952]: raise self.value [ 607.338915] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.338915] env[62952]: updated_port = self._update_port( [ 607.338915] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.338915] env[62952]: _ensure_no_port_binding_failure(port) [ 607.338915] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.338915] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 607.339624] env[62952]: nova.exception.PortBindingFailed: Binding failed for port f2e7e01f-0ace-4b09-a54a-cbc7a33c5f3f, please check neutron logs for more information. [ 607.339624] env[62952]: Removing descriptor: 17 [ 607.344212] env[62952]: ERROR nova.compute.manager [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f2e7e01f-0ace-4b09-a54a-cbc7a33c5f3f, please check neutron logs for more information. [ 607.344212] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Traceback (most recent call last): [ 607.344212] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 607.344212] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] yield resources [ 607.344212] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 607.344212] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] self.driver.spawn(context, instance, image_meta, [ 607.344212] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 607.344212] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.344212] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.344212] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] vm_ref = self.build_virtual_machine(instance, [ 607.344212] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.344544] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.344544] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.344544] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] for vif in network_info: [ 607.344544] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.344544] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] return self._sync_wrapper(fn, *args, **kwargs) [ 607.344544] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.344544] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] self.wait() [ 607.344544] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.344544] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] self[:] = self._gt.wait() [ 607.344544] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.344544] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] return self._exit_event.wait() [ 607.344544] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 607.344544] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] result = hub.switch() [ 607.344868] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 607.344868] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] return self.greenlet.switch() [ 607.344868] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.344868] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] result = function(*args, **kwargs) [ 607.344868] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.344868] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] return func(*args, **kwargs) [ 607.344868] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.344868] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] raise e [ 607.344868] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.344868] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] nwinfo = self.network_api.allocate_for_instance( [ 607.344868] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.344868] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] created_port_ids = self._update_ports_for_instance( [ 607.344868] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.345293] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] with excutils.save_and_reraise_exception(): [ 607.345293] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.345293] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] self.force_reraise() [ 607.345293] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.345293] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] raise self.value [ 607.345293] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.345293] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] updated_port = self._update_port( [ 607.345293] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.345293] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] _ensure_no_port_binding_failure(port) [ 607.345293] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.345293] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] raise exception.PortBindingFailed(port_id=port['id']) [ 607.345293] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] nova.exception.PortBindingFailed: Binding failed for port f2e7e01f-0ace-4b09-a54a-cbc7a33c5f3f, please check neutron logs for more information. [ 607.345293] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] [ 607.345611] env[62952]: INFO nova.compute.manager [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Terminating instance [ 607.345611] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Acquiring lock "refresh_cache-20cad4d8-5a23-47be-8603-2bf576700a14" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.345611] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Acquired lock "refresh_cache-20cad4d8-5a23-47be-8603-2bf576700a14" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.345761] env[62952]: DEBUG nova.network.neutron [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 607.491920] env[62952]: DEBUG nova.network.neutron [-] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.539717] env[62952]: DEBUG nova.compute.manager [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 607.549386] env[62952]: DEBUG nova.compute.manager [req-1aa2125a-4479-4883-b5e0-2c7b9ab76541 req-59874e67-b7eb-4964-91fd-df09263099e5 service nova] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Received event network-vif-deleted-6a9a4e0e-736f-4156-8381-eb26aaf6028c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 607.806495] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d46808-0f82-47b6-8823-e200bbdce7a7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.812749] env[62952]: DEBUG nova.compute.manager [req-c69fb1ae-3206-4f23-94a4-c43f53ffb64c req-3e951c45-9c17-430d-be81-5d03865d305c service nova] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Received event network-vif-deleted-5f1fb2a9-4d00-45a4-b6f3-6a9766d92324 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 607.819953] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a77b049-83df-419d-843e-945b69209ef5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.861543] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd4be70-17d3-4981-934a-1e8319c05ca9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.870016] env[62952]: DEBUG oslo_concurrency.lockutils [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Acquiring lock "6c87dd22-6c47-461a-8129-585d1cd72a2a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.870384] env[62952]: DEBUG oslo_concurrency.lockutils [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Lock "6c87dd22-6c47-461a-8129-585d1cd72a2a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.870577] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Acquiring lock "c377d16d-2641-483f-892b-42f6f067e9bc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.870755] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Lock "c377d16d-2641-483f-892b-42f6f067e9bc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.877041] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db451d03-1cb2-41ee-baf9-9067470a80ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.894381] env[62952]: DEBUG nova.compute.provider_tree [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.919062] env[62952]: DEBUG nova.network.neutron [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.999018] env[62952]: INFO nova.compute.manager [-] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Took 1.06 seconds to deallocate network for instance. [ 607.999905] env[62952]: DEBUG nova.compute.claims [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 608.000009] env[62952]: DEBUG oslo_concurrency.lockutils [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.115782] env[62952]: DEBUG nova.network.neutron [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.129174] env[62952]: DEBUG nova.network.neutron [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Successfully created port: 2f040d58-ad7e-41c5-a6c7-27bc621f5b0d {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 608.398575] env[62952]: DEBUG nova.scheduler.client.report [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.551988] env[62952]: DEBUG nova.compute.manager [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 608.593656] env[62952]: DEBUG nova.virt.hardware [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 608.593656] env[62952]: DEBUG nova.virt.hardware [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 608.594531] env[62952]: DEBUG nova.virt.hardware [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 608.594531] env[62952]: DEBUG nova.virt.hardware [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 608.597467] env[62952]: DEBUG nova.virt.hardware [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 608.597467] env[62952]: DEBUG nova.virt.hardware [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 608.597467] env[62952]: DEBUG nova.virt.hardware [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 608.597467] env[62952]: DEBUG nova.virt.hardware [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 608.597467] env[62952]: DEBUG nova.virt.hardware [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 608.597863] env[62952]: DEBUG nova.virt.hardware [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 608.597863] env[62952]: DEBUG nova.virt.hardware [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 608.598399] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7b472b-76ec-4689-9d74-fb73f12dbead {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.608993] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31bbbac1-7420-4fa8-998a-d2c80b19c859 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.626611] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Releasing lock "refresh_cache-20cad4d8-5a23-47be-8603-2bf576700a14" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.627063] env[62952]: DEBUG nova.compute.manager [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 608.627256] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 608.627762] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-29b53aea-8ee0-41d0-ae4c-7ed8dccaa934 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.639229] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b272013-17bf-4aba-adca-db38c12637b7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.663908] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 20cad4d8-5a23-47be-8603-2bf576700a14 could not be found. [ 608.664250] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 608.664424] env[62952]: INFO nova.compute.manager [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Took 0.04 seconds to destroy the instance on the hypervisor. [ 608.665851] env[62952]: DEBUG oslo.service.loopingcall [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 608.665851] env[62952]: DEBUG nova.compute.manager [-] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.665851] env[62952]: DEBUG nova.network.neutron [-] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 608.710494] env[62952]: DEBUG nova.network.neutron [-] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.753089] env[62952]: DEBUG oslo_concurrency.lockutils [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Acquiring lock "25e10d02-8956-4d53-a1a4-b977da38f1c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.753502] env[62952]: DEBUG oslo_concurrency.lockutils [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Lock "25e10d02-8956-4d53-a1a4-b977da38f1c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.873051] env[62952]: ERROR nova.compute.manager [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0c947d5e-7083-46d4-9b46-6e4a7331b432, please check neutron logs for more information. [ 608.873051] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 608.873051] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.873051] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 608.873051] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.873051] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 608.873051] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.873051] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 608.873051] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.873051] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 608.873051] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.873051] env[62952]: ERROR nova.compute.manager raise self.value [ 608.873051] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.873051] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 608.873051] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.873051] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 608.873507] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.873507] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 608.873507] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0c947d5e-7083-46d4-9b46-6e4a7331b432, please check neutron logs for more information. [ 608.873507] env[62952]: ERROR nova.compute.manager [ 608.873507] env[62952]: Traceback (most recent call last): [ 608.873507] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 608.873507] env[62952]: listener.cb(fileno) [ 608.873507] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.873507] env[62952]: result = function(*args, **kwargs) [ 608.873507] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.873507] env[62952]: return func(*args, **kwargs) [ 608.873507] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.873507] env[62952]: raise e [ 608.873507] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.873507] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 608.873507] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.873507] env[62952]: created_port_ids = self._update_ports_for_instance( [ 608.873507] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.873507] env[62952]: with excutils.save_and_reraise_exception(): [ 608.873507] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.873507] env[62952]: self.force_reraise() [ 608.873507] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.873507] env[62952]: raise self.value [ 608.873507] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.873507] env[62952]: updated_port = self._update_port( [ 608.873507] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.873507] env[62952]: _ensure_no_port_binding_failure(port) [ 608.873507] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.873507] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 608.874482] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 0c947d5e-7083-46d4-9b46-6e4a7331b432, please check neutron logs for more information. [ 608.874482] env[62952]: Removing descriptor: 18 [ 608.874482] env[62952]: ERROR nova.compute.manager [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0c947d5e-7083-46d4-9b46-6e4a7331b432, please check neutron logs for more information. [ 608.874482] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Traceback (most recent call last): [ 608.874482] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 608.874482] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] yield resources [ 608.874482] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 608.874482] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] self.driver.spawn(context, instance, image_meta, [ 608.874482] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 608.874482] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] self._vmops.spawn(context, instance, image_meta, injected_files, [ 608.874482] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 608.874482] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] vm_ref = self.build_virtual_machine(instance, [ 608.874814] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 608.874814] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] vif_infos = vmwarevif.get_vif_info(self._session, [ 608.874814] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 608.874814] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] for vif in network_info: [ 608.874814] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 608.874814] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] return self._sync_wrapper(fn, *args, **kwargs) [ 608.874814] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 608.874814] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] self.wait() [ 608.874814] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 608.874814] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] self[:] = self._gt.wait() [ 608.874814] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 608.874814] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] return self._exit_event.wait() [ 608.874814] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 608.875194] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] result = hub.switch() [ 608.875194] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 608.875194] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] return self.greenlet.switch() [ 608.875194] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.875194] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] result = function(*args, **kwargs) [ 608.875194] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.875194] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] return func(*args, **kwargs) [ 608.875194] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.875194] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] raise e [ 608.875194] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.875194] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] nwinfo = self.network_api.allocate_for_instance( [ 608.875194] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.875194] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] created_port_ids = self._update_ports_for_instance( [ 608.876100] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.876100] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] with excutils.save_and_reraise_exception(): [ 608.876100] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.876100] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] self.force_reraise() [ 608.876100] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.876100] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] raise self.value [ 608.876100] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.876100] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] updated_port = self._update_port( [ 608.876100] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.876100] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] _ensure_no_port_binding_failure(port) [ 608.876100] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.876100] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] raise exception.PortBindingFailed(port_id=port['id']) [ 608.876468] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] nova.exception.PortBindingFailed: Binding failed for port 0c947d5e-7083-46d4-9b46-6e4a7331b432, please check neutron logs for more information. [ 608.876468] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] [ 608.876468] env[62952]: INFO nova.compute.manager [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Terminating instance [ 608.876468] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Acquiring lock "refresh_cache-0e49e1b2-076a-4770-bbdc-3d5f7cf67590" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.876468] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Acquired lock "refresh_cache-0e49e1b2-076a-4770-bbdc-3d5f7cf67590" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.876664] env[62952]: DEBUG nova.network.neutron [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 608.906611] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.377s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.906611] env[62952]: DEBUG nova.compute.manager [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 608.909392] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.870s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.910967] env[62952]: INFO nova.compute.claims [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 609.213395] env[62952]: DEBUG nova.network.neutron [-] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.420655] env[62952]: DEBUG nova.compute.utils [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 609.424987] env[62952]: DEBUG nova.compute.manager [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 609.424987] env[62952]: DEBUG nova.network.neutron [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 609.428087] env[62952]: DEBUG nova.network.neutron [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.594505] env[62952]: DEBUG nova.policy [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f5d886a6fd848038f205a20c89004ed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '486a533a51ef4b0e8afdebfd56db0d9c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 609.647286] env[62952]: DEBUG nova.network.neutron [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.681799] env[62952]: ERROR nova.compute.manager [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d5a318e4-4330-4101-a6d4-803e4a6f63ae, please check neutron logs for more information. [ 609.681799] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 609.681799] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.681799] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 609.681799] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.681799] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 609.681799] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.681799] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 609.681799] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.681799] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 609.681799] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.681799] env[62952]: ERROR nova.compute.manager raise self.value [ 609.681799] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.681799] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 609.681799] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.681799] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 609.682229] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.682229] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 609.682229] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d5a318e4-4330-4101-a6d4-803e4a6f63ae, please check neutron logs for more information. [ 609.682229] env[62952]: ERROR nova.compute.manager [ 609.682229] env[62952]: Traceback (most recent call last): [ 609.682229] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 609.682229] env[62952]: listener.cb(fileno) [ 609.682229] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.682229] env[62952]: result = function(*args, **kwargs) [ 609.682229] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.682229] env[62952]: return func(*args, **kwargs) [ 609.682229] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.682229] env[62952]: raise e [ 609.682229] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.682229] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 609.682229] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.682229] env[62952]: created_port_ids = self._update_ports_for_instance( [ 609.682229] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.682229] env[62952]: with excutils.save_and_reraise_exception(): [ 609.682229] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.682229] env[62952]: self.force_reraise() [ 609.682229] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.682229] env[62952]: raise self.value [ 609.682229] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.682229] env[62952]: updated_port = self._update_port( [ 609.682229] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.682229] env[62952]: _ensure_no_port_binding_failure(port) [ 609.682229] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.682229] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 609.683203] env[62952]: nova.exception.PortBindingFailed: Binding failed for port d5a318e4-4330-4101-a6d4-803e4a6f63ae, please check neutron logs for more information. [ 609.683203] env[62952]: Removing descriptor: 15 [ 609.683203] env[62952]: ERROR nova.compute.manager [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d5a318e4-4330-4101-a6d4-803e4a6f63ae, please check neutron logs for more information. [ 609.683203] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Traceback (most recent call last): [ 609.683203] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 609.683203] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] yield resources [ 609.683203] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.683203] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] self.driver.spawn(context, instance, image_meta, [ 609.683203] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 609.683203] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.683203] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.683203] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] vm_ref = self.build_virtual_machine(instance, [ 609.683550] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.683550] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.683550] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.683550] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] for vif in network_info: [ 609.683550] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.683550] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] return self._sync_wrapper(fn, *args, **kwargs) [ 609.683550] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.683550] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] self.wait() [ 609.683550] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.683550] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] self[:] = self._gt.wait() [ 609.683550] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.683550] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] return self._exit_event.wait() [ 609.683550] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.684035] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] result = hub.switch() [ 609.684035] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.684035] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] return self.greenlet.switch() [ 609.684035] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.684035] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] result = function(*args, **kwargs) [ 609.684035] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.684035] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] return func(*args, **kwargs) [ 609.684035] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.684035] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] raise e [ 609.684035] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.684035] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] nwinfo = self.network_api.allocate_for_instance( [ 609.684035] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.684035] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] created_port_ids = self._update_ports_for_instance( [ 609.684657] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.684657] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] with excutils.save_and_reraise_exception(): [ 609.684657] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.684657] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] self.force_reraise() [ 609.684657] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.684657] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] raise self.value [ 609.684657] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.684657] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] updated_port = self._update_port( [ 609.684657] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.684657] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] _ensure_no_port_binding_failure(port) [ 609.684657] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.684657] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] raise exception.PortBindingFailed(port_id=port['id']) [ 609.684966] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] nova.exception.PortBindingFailed: Binding failed for port d5a318e4-4330-4101-a6d4-803e4a6f63ae, please check neutron logs for more information. [ 609.684966] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] [ 609.684966] env[62952]: INFO nova.compute.manager [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Terminating instance [ 609.686921] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Acquiring lock "refresh_cache-4c40c507-b988-498c-b68a-a5304541a1c4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.686921] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Acquired lock "refresh_cache-4c40c507-b988-498c-b68a-a5304541a1c4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.686921] env[62952]: DEBUG nova.network.neutron [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 609.716564] env[62952]: INFO nova.compute.manager [-] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Took 1.05 seconds to deallocate network for instance. [ 609.727469] env[62952]: DEBUG nova.compute.claims [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 609.727469] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.925056] env[62952]: DEBUG nova.compute.manager [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 610.157033] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Releasing lock "refresh_cache-0e49e1b2-076a-4770-bbdc-3d5f7cf67590" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.157480] env[62952]: DEBUG nova.compute.manager [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 610.158979] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 610.160138] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-92f3e2fd-e274-48cd-9ed7-863229adefca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.175273] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7c1017-e088-427a-8389-809102c25653 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.191806] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff98210-e1d7-4802-a010-15bc8b5aeab5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.207682] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0e49e1b2-076a-4770-bbdc-3d5f7cf67590 could not be found. [ 610.207760] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 610.207906] env[62952]: INFO nova.compute.manager [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Took 0.05 seconds to destroy the instance on the hypervisor. [ 610.208166] env[62952]: DEBUG oslo.service.loopingcall [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 610.211022] env[62952]: DEBUG nova.compute.manager [-] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.211022] env[62952]: DEBUG nova.network.neutron [-] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 610.212859] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5773b518-af85-4497-b4cf-169af5f01227 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.245479] env[62952]: DEBUG nova.network.neutron [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.248097] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f6ead81-3a96-47f4-abcd-f95b605af6d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.260398] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3acadb54-ec97-41ec-9153-c4869b7fd8f3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.268040] env[62952]: DEBUG nova.network.neutron [-] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.279990] env[62952]: DEBUG nova.compute.provider_tree [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.419529] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Acquiring lock "78a0acf9-12f5-4699-a4b8-2cbefcbd343b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.419875] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Lock "78a0acf9-12f5-4699-a4b8-2cbefcbd343b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.447912] env[62952]: DEBUG nova.network.neutron [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.715747] env[62952]: DEBUG nova.network.neutron [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Successfully created port: 15b94fb4-8441-4c4d-bcbd-b444f31b2856 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 610.770770] env[62952]: DEBUG nova.network.neutron [-] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.783903] env[62952]: DEBUG nova.scheduler.client.report [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 610.940836] env[62952]: DEBUG nova.compute.manager [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 610.951913] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Releasing lock "refresh_cache-4c40c507-b988-498c-b68a-a5304541a1c4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.951913] env[62952]: DEBUG nova.compute.manager [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 610.951913] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 610.952200] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7768b8b3-ef47-4d30-8315-80a5b4afd131 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.965651] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e400529f-8f0a-4679-8077-51e5c37546f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.994693] env[62952]: DEBUG nova.virt.hardware [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 610.995263] env[62952]: DEBUG nova.virt.hardware [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 610.995263] env[62952]: DEBUG nova.virt.hardware [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.995263] env[62952]: DEBUG nova.virt.hardware [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 610.995469] env[62952]: DEBUG nova.virt.hardware [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.995862] env[62952]: DEBUG nova.virt.hardware [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 610.995862] env[62952]: DEBUG nova.virt.hardware [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 610.995862] env[62952]: DEBUG nova.virt.hardware [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 610.996105] env[62952]: DEBUG nova.virt.hardware [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 610.996302] env[62952]: DEBUG nova.virt.hardware [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 610.996362] env[62952]: DEBUG nova.virt.hardware [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.999983] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-172ad9b9-62c5-4195-818d-8de84b65cc1a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.003124] env[62952]: DEBUG oslo_concurrency.lockutils [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Acquiring lock "4eb9ea32-b2eb-440c-871b-6942590d9c01" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.003265] env[62952]: DEBUG oslo_concurrency.lockutils [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Lock "4eb9ea32-b2eb-440c-871b-6942590d9c01" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.009364] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53cba5f7-4e4d-4525-b051-b5f607270484 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.017266] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4c40c507-b988-498c-b68a-a5304541a1c4 could not be found. [ 611.017525] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 611.017644] env[62952]: INFO nova.compute.manager [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Took 0.07 seconds to destroy the instance on the hypervisor. [ 611.017866] env[62952]: DEBUG oslo.service.loopingcall [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 611.018479] env[62952]: DEBUG nova.compute.manager [-] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.018604] env[62952]: DEBUG nova.network.neutron [-] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 611.034421] env[62952]: DEBUG nova.compute.manager [req-06952648-cfc0-46dc-90fb-e226ddab26a8 req-b4ffb608-8f1a-4d5e-9018-683cae2f16e4 service nova] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Received event network-changed-0c947d5e-7083-46d4-9b46-6e4a7331b432 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 611.034421] env[62952]: DEBUG nova.compute.manager [req-06952648-cfc0-46dc-90fb-e226ddab26a8 req-b4ffb608-8f1a-4d5e-9018-683cae2f16e4 service nova] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Refreshing instance network info cache due to event network-changed-0c947d5e-7083-46d4-9b46-6e4a7331b432. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 611.034421] env[62952]: DEBUG oslo_concurrency.lockutils [req-06952648-cfc0-46dc-90fb-e226ddab26a8 req-b4ffb608-8f1a-4d5e-9018-683cae2f16e4 service nova] Acquiring lock "refresh_cache-0e49e1b2-076a-4770-bbdc-3d5f7cf67590" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.034421] env[62952]: DEBUG oslo_concurrency.lockutils [req-06952648-cfc0-46dc-90fb-e226ddab26a8 req-b4ffb608-8f1a-4d5e-9018-683cae2f16e4 service nova] Acquired lock "refresh_cache-0e49e1b2-076a-4770-bbdc-3d5f7cf67590" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.034421] env[62952]: DEBUG nova.network.neutron [req-06952648-cfc0-46dc-90fb-e226ddab26a8 req-b4ffb608-8f1a-4d5e-9018-683cae2f16e4 service nova] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Refreshing network info cache for port 0c947d5e-7083-46d4-9b46-6e4a7331b432 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 611.076762] env[62952]: DEBUG nova.network.neutron [-] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.124939] env[62952]: DEBUG nova.compute.manager [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Received event network-changed-f2e7e01f-0ace-4b09-a54a-cbc7a33c5f3f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 611.124939] env[62952]: DEBUG nova.compute.manager [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Refreshing instance network info cache due to event network-changed-f2e7e01f-0ace-4b09-a54a-cbc7a33c5f3f. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 611.124939] env[62952]: DEBUG oslo_concurrency.lockutils [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] Acquiring lock "refresh_cache-20cad4d8-5a23-47be-8603-2bf576700a14" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.124939] env[62952]: DEBUG oslo_concurrency.lockutils [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] Acquired lock "refresh_cache-20cad4d8-5a23-47be-8603-2bf576700a14" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.124939] env[62952]: DEBUG nova.network.neutron [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Refreshing network info cache for port f2e7e01f-0ace-4b09-a54a-cbc7a33c5f3f {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 611.277010] env[62952]: INFO nova.compute.manager [-] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Took 1.07 seconds to deallocate network for instance. [ 611.284272] env[62952]: DEBUG nova.compute.claims [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 611.284272] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.288077] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.379s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.288568] env[62952]: DEBUG nova.compute.manager [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 611.292057] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.129s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.580204] env[62952]: DEBUG nova.network.neutron [-] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.593511] env[62952]: DEBUG nova.network.neutron [req-06952648-cfc0-46dc-90fb-e226ddab26a8 req-b4ffb608-8f1a-4d5e-9018-683cae2f16e4 service nova] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.602419] env[62952]: ERROR nova.compute.manager [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2f040d58-ad7e-41c5-a6c7-27bc621f5b0d, please check neutron logs for more information. [ 611.602419] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 611.602419] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.602419] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 611.602419] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 611.602419] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 611.602419] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 611.602419] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 611.602419] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.602419] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 611.602419] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.602419] env[62952]: ERROR nova.compute.manager raise self.value [ 611.602419] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 611.602419] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 611.602419] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.602419] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 611.603628] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.603628] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 611.603628] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2f040d58-ad7e-41c5-a6c7-27bc621f5b0d, please check neutron logs for more information. [ 611.603628] env[62952]: ERROR nova.compute.manager [ 611.603628] env[62952]: Traceback (most recent call last): [ 611.603628] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 611.603628] env[62952]: listener.cb(fileno) [ 611.603628] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.603628] env[62952]: result = function(*args, **kwargs) [ 611.603628] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.603628] env[62952]: return func(*args, **kwargs) [ 611.603628] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.603628] env[62952]: raise e [ 611.603628] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.603628] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 611.603628] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 611.603628] env[62952]: created_port_ids = self._update_ports_for_instance( [ 611.603628] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 611.603628] env[62952]: with excutils.save_and_reraise_exception(): [ 611.603628] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.603628] env[62952]: self.force_reraise() [ 611.603628] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.603628] env[62952]: raise self.value [ 611.603628] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 611.603628] env[62952]: updated_port = self._update_port( [ 611.603628] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.603628] env[62952]: _ensure_no_port_binding_failure(port) [ 611.603628] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.603628] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 611.604503] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 2f040d58-ad7e-41c5-a6c7-27bc621f5b0d, please check neutron logs for more information. [ 611.604503] env[62952]: Removing descriptor: 19 [ 611.604503] env[62952]: ERROR nova.compute.manager [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2f040d58-ad7e-41c5-a6c7-27bc621f5b0d, please check neutron logs for more information. [ 611.604503] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Traceback (most recent call last): [ 611.604503] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 611.604503] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] yield resources [ 611.604503] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.604503] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] self.driver.spawn(context, instance, image_meta, [ 611.604503] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 611.604503] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.604503] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.604503] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] vm_ref = self.build_virtual_machine(instance, [ 611.604857] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.604857] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.604857] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.604857] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] for vif in network_info: [ 611.604857] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.604857] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] return self._sync_wrapper(fn, *args, **kwargs) [ 611.604857] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.604857] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] self.wait() [ 611.604857] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.604857] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] self[:] = self._gt.wait() [ 611.604857] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.604857] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] return self._exit_event.wait() [ 611.604857] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.605381] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] result = hub.switch() [ 611.605381] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.605381] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] return self.greenlet.switch() [ 611.605381] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.605381] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] result = function(*args, **kwargs) [ 611.605381] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.605381] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] return func(*args, **kwargs) [ 611.605381] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.605381] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] raise e [ 611.605381] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.605381] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] nwinfo = self.network_api.allocate_for_instance( [ 611.605381] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 611.605381] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] created_port_ids = self._update_ports_for_instance( [ 611.605913] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 611.605913] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] with excutils.save_and_reraise_exception(): [ 611.605913] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.605913] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] self.force_reraise() [ 611.605913] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.605913] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] raise self.value [ 611.605913] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 611.605913] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] updated_port = self._update_port( [ 611.605913] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.605913] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] _ensure_no_port_binding_failure(port) [ 611.605913] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.605913] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] raise exception.PortBindingFailed(port_id=port['id']) [ 611.610614] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] nova.exception.PortBindingFailed: Binding failed for port 2f040d58-ad7e-41c5-a6c7-27bc621f5b0d, please check neutron logs for more information. [ 611.610614] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] [ 611.610614] env[62952]: INFO nova.compute.manager [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Terminating instance [ 611.610614] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Acquiring lock "refresh_cache-1f51dd77-519f-4d66-9048-5f63e85e3837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.610614] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Acquired lock "refresh_cache-1f51dd77-519f-4d66-9048-5f63e85e3837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.610614] env[62952]: DEBUG nova.network.neutron [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 611.675868] env[62952]: DEBUG nova.network.neutron [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.774606] env[62952]: DEBUG nova.network.neutron [req-06952648-cfc0-46dc-90fb-e226ddab26a8 req-b4ffb608-8f1a-4d5e-9018-683cae2f16e4 service nova] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.798575] env[62952]: DEBUG nova.compute.utils [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 611.809869] env[62952]: DEBUG nova.compute.manager [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 611.809869] env[62952]: DEBUG nova.network.neutron [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 611.836137] env[62952]: DEBUG nova.network.neutron [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.904686] env[62952]: DEBUG nova.policy [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e5b3cae6527f4f0db0794bbca1c5b6ac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ec2340d961b24ec9846d0326186bae6f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 612.086445] env[62952]: INFO nova.compute.manager [-] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Took 1.07 seconds to deallocate network for instance. [ 612.090049] env[62952]: DEBUG nova.compute.claims [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 612.090049] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.119798] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c7690db-9272-4cbd-985e-e673d240d511 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.127628] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b22c0ce-3ba3-418c-bd02-66a69d768238 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.168982] env[62952]: DEBUG nova.network.neutron [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.169571] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0576dfb3-340a-409e-95e7-a24cbb45a4c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.178027] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcbf5431-fc83-4ac5-9509-12ece9609c16 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.194511] env[62952]: DEBUG nova.compute.provider_tree [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.280219] env[62952]: DEBUG oslo_concurrency.lockutils [req-06952648-cfc0-46dc-90fb-e226ddab26a8 req-b4ffb608-8f1a-4d5e-9018-683cae2f16e4 service nova] Releasing lock "refresh_cache-0e49e1b2-076a-4770-bbdc-3d5f7cf67590" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.280499] env[62952]: DEBUG nova.compute.manager [req-06952648-cfc0-46dc-90fb-e226ddab26a8 req-b4ffb608-8f1a-4d5e-9018-683cae2f16e4 service nova] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Received event network-vif-deleted-0c947d5e-7083-46d4-9b46-6e4a7331b432 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 612.314660] env[62952]: DEBUG nova.compute.manager [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 612.339344] env[62952]: DEBUG oslo_concurrency.lockutils [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] Releasing lock "refresh_cache-20cad4d8-5a23-47be-8603-2bf576700a14" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.339615] env[62952]: DEBUG nova.compute.manager [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Received event network-vif-deleted-f2e7e01f-0ace-4b09-a54a-cbc7a33c5f3f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 612.339814] env[62952]: DEBUG nova.compute.manager [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Received event network-changed-d5a318e4-4330-4101-a6d4-803e4a6f63ae {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 612.339932] env[62952]: DEBUG nova.compute.manager [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Refreshing instance network info cache due to event network-changed-d5a318e4-4330-4101-a6d4-803e4a6f63ae. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 612.340157] env[62952]: DEBUG oslo_concurrency.lockutils [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] Acquiring lock "refresh_cache-4c40c507-b988-498c-b68a-a5304541a1c4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.340287] env[62952]: DEBUG oslo_concurrency.lockutils [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] Acquired lock "refresh_cache-4c40c507-b988-498c-b68a-a5304541a1c4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.340439] env[62952]: DEBUG nova.network.neutron [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Refreshing network info cache for port d5a318e4-4330-4101-a6d4-803e4a6f63ae {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 612.412908] env[62952]: DEBUG nova.network.neutron [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.544142] env[62952]: DEBUG nova.network.neutron [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Successfully created port: b18d52c0-132e-44f5-a200-b977563b8c54 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 612.698063] env[62952]: DEBUG nova.scheduler.client.report [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 612.895685] env[62952]: DEBUG nova.network.neutron [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.920984] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Releasing lock "refresh_cache-1f51dd77-519f-4d66-9048-5f63e85e3837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.921125] env[62952]: DEBUG nova.compute.manager [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 612.921634] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 612.921634] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a4bad29d-6db6-4bb9-8f5b-96e60b72132c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.936152] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8694d2b-5e6c-4b1e-b40e-54e3f1e51405 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.967305] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1f51dd77-519f-4d66-9048-5f63e85e3837 could not be found. [ 612.967572] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 612.967716] env[62952]: INFO nova.compute.manager [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Took 0.05 seconds to destroy the instance on the hypervisor. [ 612.967931] env[62952]: DEBUG oslo.service.loopingcall [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.968178] env[62952]: DEBUG nova.compute.manager [-] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.968278] env[62952]: DEBUG nova.network.neutron [-] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 613.086032] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Acquiring lock "f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.086276] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Lock "f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.141622] env[62952]: DEBUG nova.network.neutron [-] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 613.167275] env[62952]: DEBUG nova.network.neutron [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.204294] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.913s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.205351] env[62952]: ERROR nova.compute.manager [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 91d1c070-8a67-4af0-9df8-aa8c642d8552, please check neutron logs for more information. [ 613.205351] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Traceback (most recent call last): [ 613.205351] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 613.205351] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] self.driver.spawn(context, instance, image_meta, [ 613.205351] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 613.205351] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.205351] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.205351] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] vm_ref = self.build_virtual_machine(instance, [ 613.205351] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.205351] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.205351] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.205649] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] for vif in network_info: [ 613.205649] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.205649] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] return self._sync_wrapper(fn, *args, **kwargs) [ 613.205649] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.205649] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] self.wait() [ 613.205649] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.205649] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] self[:] = self._gt.wait() [ 613.205649] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.205649] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] return self._exit_event.wait() [ 613.205649] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 613.205649] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] result = hub.switch() [ 613.205649] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 613.205649] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] return self.greenlet.switch() [ 613.206013] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.206013] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] result = function(*args, **kwargs) [ 613.206013] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 613.206013] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] return func(*args, **kwargs) [ 613.206013] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.206013] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] raise e [ 613.206013] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.206013] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] nwinfo = self.network_api.allocate_for_instance( [ 613.206013] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.206013] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] created_port_ids = self._update_ports_for_instance( [ 613.206013] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.206013] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] with excutils.save_and_reraise_exception(): [ 613.206013] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.206398] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] self.force_reraise() [ 613.206398] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.206398] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] raise self.value [ 613.206398] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.206398] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] updated_port = self._update_port( [ 613.206398] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.206398] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] _ensure_no_port_binding_failure(port) [ 613.206398] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.206398] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] raise exception.PortBindingFailed(port_id=port['id']) [ 613.206398] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] nova.exception.PortBindingFailed: Binding failed for port 91d1c070-8a67-4af0-9df8-aa8c642d8552, please check neutron logs for more information. [ 613.206398] env[62952]: ERROR nova.compute.manager [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] [ 613.208189] env[62952]: DEBUG nova.compute.utils [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Binding failed for port 91d1c070-8a67-4af0-9df8-aa8c642d8552, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 613.208189] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 8.819s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.208189] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.213814] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 613.214187] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 7.721s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.223446] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81211689-d8c0-49ac-9f64-e73fdd067f15 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.231499] env[62952]: DEBUG nova.compute.manager [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Build of instance d19a78e1-ac0d-4b3e-9812-d6938298a255 was re-scheduled: Binding failed for port 91d1c070-8a67-4af0-9df8-aa8c642d8552, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 613.231499] env[62952]: DEBUG nova.compute.manager [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 613.231499] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "refresh_cache-d19a78e1-ac0d-4b3e-9812-d6938298a255" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.231499] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquired lock "refresh_cache-d19a78e1-ac0d-4b3e-9812-d6938298a255" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.231499] env[62952]: DEBUG nova.network.neutron [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 613.234614] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232beb6b-d247-486f-b945-fff62dada198 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.252250] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c547f286-8c85-45bc-b17e-9f237ba302d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.263270] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec3313a4-8a98-4e09-aada-949de1d46188 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.302035] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181552MB free_disk=145GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 613.302035] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.326651] env[62952]: DEBUG nova.compute.manager [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 613.356979] env[62952]: DEBUG nova.virt.hardware [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 613.357137] env[62952]: DEBUG nova.virt.hardware [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 613.357245] env[62952]: DEBUG nova.virt.hardware [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 613.357407] env[62952]: DEBUG nova.virt.hardware [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 613.357519] env[62952]: DEBUG nova.virt.hardware [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 613.357687] env[62952]: DEBUG nova.virt.hardware [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 613.357970] env[62952]: DEBUG nova.virt.hardware [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 613.358291] env[62952]: DEBUG nova.virt.hardware [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 613.358356] env[62952]: DEBUG nova.virt.hardware [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 613.358515] env[62952]: DEBUG nova.virt.hardware [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 613.358681] env[62952]: DEBUG nova.virt.hardware [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 613.359584] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9e86905-b9b9-4a50-ac31-0ede72c7e27c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.367730] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56ac9f8-f5f3-4053-830b-7c1ef9131b31 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.646243] env[62952]: DEBUG nova.network.neutron [-] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.672511] env[62952]: DEBUG oslo_concurrency.lockutils [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] Releasing lock "refresh_cache-4c40c507-b988-498c-b68a-a5304541a1c4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.672768] env[62952]: DEBUG nova.compute.manager [req-9d2ad18b-1ade-4f91-8dbf-a76382d3d71d req-123d841c-3bfc-4ac0-af08-db6cf1b72b19 service nova] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Received event network-vif-deleted-d5a318e4-4330-4101-a6d4-803e4a6f63ae {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 613.773578] env[62952]: DEBUG nova.network.neutron [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 613.978010] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a91c2c0-5548-4c6f-ab8d-5d95206b6fcc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.985677] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08872153-bcc5-4e20-83c7-f46860b9f3c9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.017370] env[62952]: DEBUG nova.network.neutron [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.021878] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f1b1397-e47c-4bfd-b2c9-4baa68b6114e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.027123] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0d2e85-a943-4cba-84f7-67c0f5b47dd5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.040753] env[62952]: DEBUG nova.compute.provider_tree [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.052398] env[62952]: ERROR nova.compute.manager [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 15b94fb4-8441-4c4d-bcbd-b444f31b2856, please check neutron logs for more information. [ 614.052398] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 614.052398] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.052398] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 614.052398] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.052398] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 614.052398] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.052398] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 614.052398] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.052398] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 614.052398] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.052398] env[62952]: ERROR nova.compute.manager raise self.value [ 614.052398] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.052398] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 614.052398] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.052398] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 614.053614] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.053614] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 614.053614] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 15b94fb4-8441-4c4d-bcbd-b444f31b2856, please check neutron logs for more information. [ 614.053614] env[62952]: ERROR nova.compute.manager [ 614.053614] env[62952]: Traceback (most recent call last): [ 614.053614] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 614.053614] env[62952]: listener.cb(fileno) [ 614.053614] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.053614] env[62952]: result = function(*args, **kwargs) [ 614.053614] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.053614] env[62952]: return func(*args, **kwargs) [ 614.053614] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.053614] env[62952]: raise e [ 614.053614] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.053614] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 614.053614] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.053614] env[62952]: created_port_ids = self._update_ports_for_instance( [ 614.053614] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.053614] env[62952]: with excutils.save_and_reraise_exception(): [ 614.053614] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.053614] env[62952]: self.force_reraise() [ 614.053614] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.053614] env[62952]: raise self.value [ 614.053614] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.053614] env[62952]: updated_port = self._update_port( [ 614.053614] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.053614] env[62952]: _ensure_no_port_binding_failure(port) [ 614.053614] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.053614] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 614.056500] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 15b94fb4-8441-4c4d-bcbd-b444f31b2856, please check neutron logs for more information. [ 614.056500] env[62952]: Removing descriptor: 17 [ 614.056500] env[62952]: ERROR nova.compute.manager [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 15b94fb4-8441-4c4d-bcbd-b444f31b2856, please check neutron logs for more information. [ 614.056500] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Traceback (most recent call last): [ 614.056500] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 614.056500] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] yield resources [ 614.056500] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.056500] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] self.driver.spawn(context, instance, image_meta, [ 614.056500] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 614.056500] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.056500] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.056500] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] vm_ref = self.build_virtual_machine(instance, [ 614.057021] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.057021] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.057021] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.057021] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] for vif in network_info: [ 614.057021] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 614.057021] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] return self._sync_wrapper(fn, *args, **kwargs) [ 614.057021] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 614.057021] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] self.wait() [ 614.057021] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 614.057021] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] self[:] = self._gt.wait() [ 614.057021] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.057021] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] return self._exit_event.wait() [ 614.057021] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 614.057353] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] result = hub.switch() [ 614.057353] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 614.057353] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] return self.greenlet.switch() [ 614.057353] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.057353] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] result = function(*args, **kwargs) [ 614.057353] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.057353] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] return func(*args, **kwargs) [ 614.057353] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.057353] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] raise e [ 614.057353] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.057353] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] nwinfo = self.network_api.allocate_for_instance( [ 614.057353] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.057353] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] created_port_ids = self._update_ports_for_instance( [ 614.057680] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.057680] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] with excutils.save_and_reraise_exception(): [ 614.057680] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.057680] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] self.force_reraise() [ 614.057680] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.057680] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] raise self.value [ 614.057680] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.057680] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] updated_port = self._update_port( [ 614.057680] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.057680] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] _ensure_no_port_binding_failure(port) [ 614.057680] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.057680] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] raise exception.PortBindingFailed(port_id=port['id']) [ 614.057969] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] nova.exception.PortBindingFailed: Binding failed for port 15b94fb4-8441-4c4d-bcbd-b444f31b2856, please check neutron logs for more information. [ 614.057969] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] [ 614.057969] env[62952]: INFO nova.compute.manager [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Terminating instance [ 614.057969] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Acquiring lock "refresh_cache-5db9f2cd-faf1-4972-8fa0-7b7705f40bf4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.057969] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Acquired lock "refresh_cache-5db9f2cd-faf1-4972-8fa0-7b7705f40bf4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.057969] env[62952]: DEBUG nova.network.neutron [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 614.149383] env[62952]: INFO nova.compute.manager [-] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Took 1.18 seconds to deallocate network for instance. [ 614.151822] env[62952]: DEBUG nova.compute.claims [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 614.151896] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.525769] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Releasing lock "refresh_cache-d19a78e1-ac0d-4b3e-9812-d6938298a255" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.525769] env[62952]: DEBUG nova.compute.manager [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 614.525769] env[62952]: DEBUG nova.compute.manager [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.525769] env[62952]: DEBUG nova.network.neutron [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 614.543471] env[62952]: DEBUG nova.scheduler.client.report [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.558482] env[62952]: DEBUG nova.network.neutron [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 614.590508] env[62952]: DEBUG nova.network.neutron [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 614.726186] env[62952]: DEBUG nova.network.neutron [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.053731] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.839s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.054431] env[62952]: ERROR nova.compute.manager [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6a9a4e0e-736f-4156-8381-eb26aaf6028c, please check neutron logs for more information. [ 615.054431] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Traceback (most recent call last): [ 615.054431] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 615.054431] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] self.driver.spawn(context, instance, image_meta, [ 615.054431] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 615.054431] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 615.054431] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 615.054431] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] vm_ref = self.build_virtual_machine(instance, [ 615.054431] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 615.054431] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] vif_infos = vmwarevif.get_vif_info(self._session, [ 615.054431] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 615.054736] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] for vif in network_info: [ 615.054736] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 615.054736] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] return self._sync_wrapper(fn, *args, **kwargs) [ 615.054736] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 615.054736] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] self.wait() [ 615.054736] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 615.054736] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] self[:] = self._gt.wait() [ 615.054736] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 615.054736] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] return self._exit_event.wait() [ 615.054736] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 615.054736] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] result = hub.switch() [ 615.054736] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 615.054736] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] return self.greenlet.switch() [ 615.055135] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.055135] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] result = function(*args, **kwargs) [ 615.055135] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.055135] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] return func(*args, **kwargs) [ 615.055135] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.055135] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] raise e [ 615.055135] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.055135] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] nwinfo = self.network_api.allocate_for_instance( [ 615.055135] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.055135] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] created_port_ids = self._update_ports_for_instance( [ 615.055135] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.055135] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] with excutils.save_and_reraise_exception(): [ 615.055135] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.055449] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] self.force_reraise() [ 615.055449] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.055449] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] raise self.value [ 615.055449] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.055449] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] updated_port = self._update_port( [ 615.055449] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.055449] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] _ensure_no_port_binding_failure(port) [ 615.055449] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.055449] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] raise exception.PortBindingFailed(port_id=port['id']) [ 615.055449] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] nova.exception.PortBindingFailed: Binding failed for port 6a9a4e0e-736f-4156-8381-eb26aaf6028c, please check neutron logs for more information. [ 615.055449] env[62952]: ERROR nova.compute.manager [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] [ 615.055756] env[62952]: DEBUG nova.compute.utils [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Binding failed for port 6a9a4e0e-736f-4156-8381-eb26aaf6028c, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 615.057904] env[62952]: ERROR nova.compute.manager [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b18d52c0-132e-44f5-a200-b977563b8c54, please check neutron logs for more information. [ 615.057904] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 615.057904] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.057904] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 615.057904] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.057904] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 615.057904] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.057904] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 615.057904] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.057904] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 615.057904] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.057904] env[62952]: ERROR nova.compute.manager raise self.value [ 615.057904] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.057904] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 615.057904] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.057904] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 615.058488] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.058488] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 615.058488] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b18d52c0-132e-44f5-a200-b977563b8c54, please check neutron logs for more information. [ 615.058488] env[62952]: ERROR nova.compute.manager [ 615.058488] env[62952]: Traceback (most recent call last): [ 615.058488] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 615.058488] env[62952]: listener.cb(fileno) [ 615.058488] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.058488] env[62952]: result = function(*args, **kwargs) [ 615.058488] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.058488] env[62952]: return func(*args, **kwargs) [ 615.058488] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.058488] env[62952]: raise e [ 615.058488] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.058488] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 615.058488] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.058488] env[62952]: created_port_ids = self._update_ports_for_instance( [ 615.058488] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.058488] env[62952]: with excutils.save_and_reraise_exception(): [ 615.058488] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.058488] env[62952]: self.force_reraise() [ 615.058488] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.058488] env[62952]: raise self.value [ 615.058488] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.058488] env[62952]: updated_port = self._update_port( [ 615.058488] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.058488] env[62952]: _ensure_no_port_binding_failure(port) [ 615.058488] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.058488] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 615.059317] env[62952]: nova.exception.PortBindingFailed: Binding failed for port b18d52c0-132e-44f5-a200-b977563b8c54, please check neutron logs for more information. [ 615.059317] env[62952]: Removing descriptor: 15 [ 615.059379] env[62952]: DEBUG nova.compute.manager [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Build of instance b419be16-cf2e-4ba5-aea3-fd2f56ff006a was re-scheduled: Binding failed for port 6a9a4e0e-736f-4156-8381-eb26aaf6028c, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 615.059834] env[62952]: DEBUG nova.compute.manager [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 615.060515] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Acquiring lock "refresh_cache-b419be16-cf2e-4ba5-aea3-fd2f56ff006a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.060797] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Acquired lock "refresh_cache-b419be16-cf2e-4ba5-aea3-fd2f56ff006a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.061080] env[62952]: DEBUG nova.network.neutron [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 615.062385] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.877s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.068655] env[62952]: INFO nova.compute.claims [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 615.071744] env[62952]: ERROR nova.compute.manager [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b18d52c0-132e-44f5-a200-b977563b8c54, please check neutron logs for more information. [ 615.071744] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Traceback (most recent call last): [ 615.071744] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 615.071744] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] yield resources [ 615.071744] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 615.071744] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] self.driver.spawn(context, instance, image_meta, [ 615.071744] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 615.071744] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] self._vmops.spawn(context, instance, image_meta, injected_files, [ 615.071744] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 615.071744] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] vm_ref = self.build_virtual_machine(instance, [ 615.071744] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 615.073353] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] vif_infos = vmwarevif.get_vif_info(self._session, [ 615.073353] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 615.073353] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] for vif in network_info: [ 615.073353] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 615.073353] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] return self._sync_wrapper(fn, *args, **kwargs) [ 615.073353] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 615.073353] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] self.wait() [ 615.073353] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 615.073353] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] self[:] = self._gt.wait() [ 615.073353] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 615.073353] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] return self._exit_event.wait() [ 615.073353] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 615.073353] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] result = hub.switch() [ 615.073938] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 615.073938] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] return self.greenlet.switch() [ 615.073938] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.073938] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] result = function(*args, **kwargs) [ 615.073938] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.073938] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] return func(*args, **kwargs) [ 615.073938] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.073938] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] raise e [ 615.073938] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.073938] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] nwinfo = self.network_api.allocate_for_instance( [ 615.073938] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.073938] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] created_port_ids = self._update_ports_for_instance( [ 615.073938] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.074327] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] with excutils.save_and_reraise_exception(): [ 615.074327] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.074327] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] self.force_reraise() [ 615.074327] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.074327] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] raise self.value [ 615.074327] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.074327] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] updated_port = self._update_port( [ 615.074327] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.074327] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] _ensure_no_port_binding_failure(port) [ 615.074327] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.074327] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] raise exception.PortBindingFailed(port_id=port['id']) [ 615.074327] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] nova.exception.PortBindingFailed: Binding failed for port b18d52c0-132e-44f5-a200-b977563b8c54, please check neutron logs for more information. [ 615.074327] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] [ 615.074683] env[62952]: INFO nova.compute.manager [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Terminating instance [ 615.074683] env[62952]: DEBUG nova.network.neutron [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.076870] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Acquiring lock "refresh_cache-dea74836-bdbb-4260-b15d-2dab04ee3940" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.076870] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Acquired lock "refresh_cache-dea74836-bdbb-4260-b15d-2dab04ee3940" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.076870] env[62952]: DEBUG nova.network.neutron [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 615.228258] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Acquiring lock "6ac19031-ac4e-4061-bbfd-75bd5c90803a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.228319] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Lock "6ac19031-ac4e-4061-bbfd-75bd5c90803a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.228645] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Releasing lock "refresh_cache-5db9f2cd-faf1-4972-8fa0-7b7705f40bf4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.228998] env[62952]: DEBUG nova.compute.manager [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 615.229239] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 615.229505] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3320053-7bff-4785-af6a-4340796ddc65 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.241197] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b566616a-603a-47e3-beec-8bca089dffea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.267052] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4 could not be found. [ 615.267361] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 615.267475] env[62952]: INFO nova.compute.manager [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 615.267710] env[62952]: DEBUG oslo.service.loopingcall [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 615.267909] env[62952]: DEBUG nova.compute.manager [-] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.267995] env[62952]: DEBUG nova.network.neutron [-] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 615.289188] env[62952]: DEBUG nova.network.neutron [-] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.294087] env[62952]: DEBUG nova.compute.manager [req-8a4e0033-4e9d-4b14-9807-ab639d5ad5c3 req-e181c63c-6fe1-482e-b7c7-8b723b434f67 service nova] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Received event network-changed-2f040d58-ad7e-41c5-a6c7-27bc621f5b0d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 615.294087] env[62952]: DEBUG nova.compute.manager [req-8a4e0033-4e9d-4b14-9807-ab639d5ad5c3 req-e181c63c-6fe1-482e-b7c7-8b723b434f67 service nova] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Refreshing instance network info cache due to event network-changed-2f040d58-ad7e-41c5-a6c7-27bc621f5b0d. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 615.294087] env[62952]: DEBUG oslo_concurrency.lockutils [req-8a4e0033-4e9d-4b14-9807-ab639d5ad5c3 req-e181c63c-6fe1-482e-b7c7-8b723b434f67 service nova] Acquiring lock "refresh_cache-1f51dd77-519f-4d66-9048-5f63e85e3837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.294087] env[62952]: DEBUG oslo_concurrency.lockutils [req-8a4e0033-4e9d-4b14-9807-ab639d5ad5c3 req-e181c63c-6fe1-482e-b7c7-8b723b434f67 service nova] Acquired lock "refresh_cache-1f51dd77-519f-4d66-9048-5f63e85e3837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.294269] env[62952]: DEBUG nova.network.neutron [req-8a4e0033-4e9d-4b14-9807-ab639d5ad5c3 req-e181c63c-6fe1-482e-b7c7-8b723b434f67 service nova] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Refreshing network info cache for port 2f040d58-ad7e-41c5-a6c7-27bc621f5b0d {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 615.582197] env[62952]: INFO nova.compute.manager [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] Took 1.06 seconds to deallocate network for instance. [ 615.730259] env[62952]: DEBUG nova.network.neutron [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.732922] env[62952]: DEBUG nova.network.neutron [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.793560] env[62952]: DEBUG nova.network.neutron [-] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.883321] env[62952]: DEBUG nova.network.neutron [req-8a4e0033-4e9d-4b14-9807-ab639d5ad5c3 req-e181c63c-6fe1-482e-b7c7-8b723b434f67 service nova] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.963573] env[62952]: DEBUG nova.network.neutron [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.072529] env[62952]: DEBUG nova.network.neutron [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.195878] env[62952]: DEBUG nova.network.neutron [req-8a4e0033-4e9d-4b14-9807-ab639d5ad5c3 req-e181c63c-6fe1-482e-b7c7-8b723b434f67 service nova] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.298516] env[62952]: INFO nova.compute.manager [-] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Took 1.03 seconds to deallocate network for instance. [ 616.301311] env[62952]: DEBUG nova.compute.claims [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 616.301605] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.392699] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84be562f-1feb-4cd3-8787-ba6f32eb75f5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.401746] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1340b613-f072-479e-8442-b1143c629582 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.437650] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2edf79ec-8930-427a-9d32-893ac59b9471 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.449027] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a5792b-b219-4704-9e5e-871b9c384027 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.463721] env[62952]: DEBUG nova.compute.provider_tree [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.470073] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Releasing lock "refresh_cache-b419be16-cf2e-4ba5-aea3-fd2f56ff006a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.470325] env[62952]: DEBUG nova.compute.manager [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 616.470553] env[62952]: DEBUG nova.compute.manager [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 616.470671] env[62952]: DEBUG nova.network.neutron [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 616.496638] env[62952]: DEBUG nova.network.neutron [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.576611] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Releasing lock "refresh_cache-dea74836-bdbb-4260-b15d-2dab04ee3940" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.579309] env[62952]: DEBUG nova.compute.manager [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 616.579309] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 616.579675] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-80df4caf-dd7f-49d2-ae32-1ccdfc4950d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.596574] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfe7d94e-7661-4e45-9442-173501164ee2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.630056] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dea74836-bdbb-4260-b15d-2dab04ee3940 could not be found. [ 616.630431] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 616.630674] env[62952]: INFO nova.compute.manager [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Took 0.05 seconds to destroy the instance on the hypervisor. [ 616.631007] env[62952]: DEBUG oslo.service.loopingcall [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 616.632277] env[62952]: DEBUG nova.compute.manager [-] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 616.632417] env[62952]: DEBUG nova.network.neutron [-] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 616.657280] env[62952]: DEBUG nova.network.neutron [-] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.666839] env[62952]: INFO nova.scheduler.client.report [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Deleted allocations for instance d19a78e1-ac0d-4b3e-9812-d6938298a255 [ 616.700570] env[62952]: DEBUG oslo_concurrency.lockutils [req-8a4e0033-4e9d-4b14-9807-ab639d5ad5c3 req-e181c63c-6fe1-482e-b7c7-8b723b434f67 service nova] Releasing lock "refresh_cache-1f51dd77-519f-4d66-9048-5f63e85e3837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.700570] env[62952]: DEBUG nova.compute.manager [req-8a4e0033-4e9d-4b14-9807-ab639d5ad5c3 req-e181c63c-6fe1-482e-b7c7-8b723b434f67 service nova] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Received event network-vif-deleted-2f040d58-ad7e-41c5-a6c7-27bc621f5b0d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 616.971074] env[62952]: DEBUG nova.scheduler.client.report [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.998321] env[62952]: DEBUG nova.network.neutron [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.159844] env[62952]: DEBUG nova.network.neutron [-] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.180653] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13e9d90f-f5c9-4684-a202-5ce001dc9e1a tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "d19a78e1-ac0d-4b3e-9812-d6938298a255" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.758s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.185678] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "d19a78e1-ac0d-4b3e-9812-d6938298a255" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 13.312s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.185678] env[62952]: INFO nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: d19a78e1-ac0d-4b3e-9812-d6938298a255] During sync_power_state the instance has a pending task (spawning). Skip. [ 617.185678] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "d19a78e1-ac0d-4b3e-9812-d6938298a255" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.477804] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.415s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.478393] env[62952]: DEBUG nova.compute.manager [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 617.487398] env[62952]: DEBUG oslo_concurrency.lockutils [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.487s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.502335] env[62952]: INFO nova.compute.manager [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] Took 1.03 seconds to deallocate network for instance. [ 617.669808] env[62952]: INFO nova.compute.manager [-] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Took 1.04 seconds to deallocate network for instance. [ 617.673310] env[62952]: DEBUG nova.compute.claims [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 617.674037] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.686203] env[62952]: DEBUG nova.compute.manager [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 617.836024] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Acquiring lock "dc025711-45be-49b6-a765-f7687b94e6a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.836024] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Lock "dc025711-45be-49b6-a765-f7687b94e6a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.999813] env[62952]: DEBUG nova.compute.utils [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 618.003836] env[62952]: DEBUG nova.compute.manager [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 618.004069] env[62952]: DEBUG nova.network.neutron [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 618.074123] env[62952]: DEBUG nova.compute.manager [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Received event network-changed-15b94fb4-8441-4c4d-bcbd-b444f31b2856 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 618.074123] env[62952]: DEBUG nova.compute.manager [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Refreshing instance network info cache due to event network-changed-15b94fb4-8441-4c4d-bcbd-b444f31b2856. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 618.074123] env[62952]: DEBUG oslo_concurrency.lockutils [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] Acquiring lock "refresh_cache-5db9f2cd-faf1-4972-8fa0-7b7705f40bf4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.074123] env[62952]: DEBUG oslo_concurrency.lockutils [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] Acquired lock "refresh_cache-5db9f2cd-faf1-4972-8fa0-7b7705f40bf4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.074560] env[62952]: DEBUG nova.network.neutron [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Refreshing network info cache for port 15b94fb4-8441-4c4d-bcbd-b444f31b2856 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 618.121819] env[62952]: DEBUG nova.policy [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b49eb44055fe4d688809e24f00b17362', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'becfc3a082164f7c91644bc037891efb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 618.212967] env[62952]: DEBUG oslo_concurrency.lockutils [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.293711] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aae458e1-4559-466d-bbb9-75b20abd0256 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.302424] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-150b3035-6a84-40e6-9ecd-5f81b1760763 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.335103] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d980c7b-1512-4f8d-9e87-7f8483bb9df1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.343856] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21781b23-eb27-4ab7-b177-602ecbfd0254 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.360743] env[62952]: DEBUG nova.compute.provider_tree [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.507523] env[62952]: DEBUG nova.compute.manager [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 618.562424] env[62952]: INFO nova.scheduler.client.report [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Deleted allocations for instance b419be16-cf2e-4ba5-aea3-fd2f56ff006a [ 618.632642] env[62952]: DEBUG nova.network.neutron [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.838803] env[62952]: DEBUG nova.network.neutron [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.863578] env[62952]: DEBUG nova.scheduler.client.report [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 619.033025] env[62952]: DEBUG nova.network.neutron [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Successfully created port: 1e4baadd-9333-4427-93c4-bad737fec2c6 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 619.073910] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce8f7f63-542c-4ec2-b4a1-5630b0cc0a04 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Lock "b419be16-cf2e-4ba5-aea3-fd2f56ff006a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.855s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.073910] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "b419be16-cf2e-4ba5-aea3-fd2f56ff006a" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 15.200s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.073910] env[62952]: INFO nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: b419be16-cf2e-4ba5-aea3-fd2f56ff006a] During sync_power_state the instance has a pending task (spawning). Skip. [ 619.073910] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "b419be16-cf2e-4ba5-aea3-fd2f56ff006a" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.310399] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Acquiring lock "0fd60040-c3ab-491f-a53e-f3297657367e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.310620] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Lock "0fd60040-c3ab-491f-a53e-f3297657367e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.343662] env[62952]: DEBUG oslo_concurrency.lockutils [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] Releasing lock "refresh_cache-5db9f2cd-faf1-4972-8fa0-7b7705f40bf4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.343930] env[62952]: DEBUG nova.compute.manager [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Received event network-vif-deleted-15b94fb4-8441-4c4d-bcbd-b444f31b2856 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 619.345036] env[62952]: DEBUG nova.compute.manager [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Received event network-changed-b18d52c0-132e-44f5-a200-b977563b8c54 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 619.345036] env[62952]: DEBUG nova.compute.manager [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Refreshing instance network info cache due to event network-changed-b18d52c0-132e-44f5-a200-b977563b8c54. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 619.345036] env[62952]: DEBUG oslo_concurrency.lockutils [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] Acquiring lock "refresh_cache-dea74836-bdbb-4260-b15d-2dab04ee3940" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.345036] env[62952]: DEBUG oslo_concurrency.lockutils [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] Acquired lock "refresh_cache-dea74836-bdbb-4260-b15d-2dab04ee3940" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.345036] env[62952]: DEBUG nova.network.neutron [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Refreshing network info cache for port b18d52c0-132e-44f5-a200-b977563b8c54 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 619.372657] env[62952]: DEBUG oslo_concurrency.lockutils [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.885s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.372960] env[62952]: ERROR nova.compute.manager [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5f1fb2a9-4d00-45a4-b6f3-6a9766d92324, please check neutron logs for more information. [ 619.372960] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Traceback (most recent call last): [ 619.372960] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 619.372960] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] self.driver.spawn(context, instance, image_meta, [ 619.372960] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 619.372960] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.372960] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.372960] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] vm_ref = self.build_virtual_machine(instance, [ 619.372960] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.372960] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.372960] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.373343] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] for vif in network_info: [ 619.373343] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.373343] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] return self._sync_wrapper(fn, *args, **kwargs) [ 619.373343] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.373343] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] self.wait() [ 619.373343] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.373343] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] self[:] = self._gt.wait() [ 619.373343] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.373343] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] return self._exit_event.wait() [ 619.373343] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 619.373343] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] result = hub.switch() [ 619.373343] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 619.373343] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] return self.greenlet.switch() [ 619.373685] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.373685] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] result = function(*args, **kwargs) [ 619.373685] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.373685] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] return func(*args, **kwargs) [ 619.373685] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.373685] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] raise e [ 619.373685] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.373685] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] nwinfo = self.network_api.allocate_for_instance( [ 619.373685] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 619.373685] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] created_port_ids = self._update_ports_for_instance( [ 619.373685] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 619.373685] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] with excutils.save_and_reraise_exception(): [ 619.373685] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.374066] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] self.force_reraise() [ 619.374066] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.374066] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] raise self.value [ 619.374066] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 619.374066] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] updated_port = self._update_port( [ 619.374066] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.374066] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] _ensure_no_port_binding_failure(port) [ 619.374066] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.374066] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] raise exception.PortBindingFailed(port_id=port['id']) [ 619.374066] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] nova.exception.PortBindingFailed: Binding failed for port 5f1fb2a9-4d00-45a4-b6f3-6a9766d92324, please check neutron logs for more information. [ 619.374066] env[62952]: ERROR nova.compute.manager [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] [ 619.374341] env[62952]: DEBUG nova.compute.utils [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Binding failed for port 5f1fb2a9-4d00-45a4-b6f3-6a9766d92324, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 619.376103] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.649s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.379561] env[62952]: DEBUG nova.compute.manager [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Build of instance 9c762648-73d9-4976-90e2-13d6a394e7f6 was re-scheduled: Binding failed for port 5f1fb2a9-4d00-45a4-b6f3-6a9766d92324, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 619.379561] env[62952]: DEBUG nova.compute.manager [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 619.379561] env[62952]: DEBUG oslo_concurrency.lockutils [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "refresh_cache-9c762648-73d9-4976-90e2-13d6a394e7f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.379561] env[62952]: DEBUG oslo_concurrency.lockutils [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "refresh_cache-9c762648-73d9-4976-90e2-13d6a394e7f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.379961] env[62952]: DEBUG nova.network.neutron [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 619.519278] env[62952]: DEBUG nova.compute.manager [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 619.546105] env[62952]: DEBUG nova.virt.hardware [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:30:50Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1140085355',id=19,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1573208181',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 619.546424] env[62952]: DEBUG nova.virt.hardware [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 619.546656] env[62952]: DEBUG nova.virt.hardware [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 619.546780] env[62952]: DEBUG nova.virt.hardware [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 619.547123] env[62952]: DEBUG nova.virt.hardware [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 619.547123] env[62952]: DEBUG nova.virt.hardware [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 619.547250] env[62952]: DEBUG nova.virt.hardware [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 619.547414] env[62952]: DEBUG nova.virt.hardware [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 619.547686] env[62952]: DEBUG nova.virt.hardware [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 619.547748] env[62952]: DEBUG nova.virt.hardware [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 619.547991] env[62952]: DEBUG nova.virt.hardware [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 619.548900] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73605b64-c9a1-4d96-bf36-0f6013a5490d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.557493] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-724c6e3b-bdfd-4a0d-9aac-074fdf22ec9a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.578196] env[62952]: DEBUG nova.compute.manager [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 619.878031] env[62952]: DEBUG nova.network.neutron [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.914775] env[62952]: DEBUG nova.network.neutron [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.108375] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.141294] env[62952]: DEBUG nova.network.neutron [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.194159] env[62952]: DEBUG nova.network.neutron [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.210584] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6eee5db-63b5-4b87-9a42-ddebee518974 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.224897] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63fb8926-e4bc-465a-abcc-8a5bf8276012 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.266822] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01983597-cc16-4116-936c-adfa342a6c1f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.275262] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-997a0d8f-2c0b-4b9a-b370-294bd2cbee69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.289859] env[62952]: DEBUG nova.compute.provider_tree [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.337747] env[62952]: ERROR nova.compute.manager [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1e4baadd-9333-4427-93c4-bad737fec2c6, please check neutron logs for more information. [ 620.337747] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 620.337747] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.337747] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 620.337747] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 620.337747] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 620.337747] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 620.337747] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 620.337747] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.337747] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 620.337747] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.337747] env[62952]: ERROR nova.compute.manager raise self.value [ 620.337747] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 620.337747] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 620.337747] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.337747] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 620.338324] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.338324] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 620.338324] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1e4baadd-9333-4427-93c4-bad737fec2c6, please check neutron logs for more information. [ 620.338324] env[62952]: ERROR nova.compute.manager [ 620.338324] env[62952]: Traceback (most recent call last): [ 620.338464] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 620.338464] env[62952]: listener.cb(fileno) [ 620.338464] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.338464] env[62952]: result = function(*args, **kwargs) [ 620.338464] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 620.338464] env[62952]: return func(*args, **kwargs) [ 620.338464] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 620.338464] env[62952]: raise e [ 620.338464] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.338464] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 620.338464] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 620.338464] env[62952]: created_port_ids = self._update_ports_for_instance( [ 620.338464] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 620.338464] env[62952]: with excutils.save_and_reraise_exception(): [ 620.338464] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.338464] env[62952]: self.force_reraise() [ 620.338464] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.338464] env[62952]: raise self.value [ 620.338464] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 620.338464] env[62952]: updated_port = self._update_port( [ 620.338464] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.338464] env[62952]: _ensure_no_port_binding_failure(port) [ 620.338464] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.338464] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 620.338464] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 1e4baadd-9333-4427-93c4-bad737fec2c6, please check neutron logs for more information. [ 620.338464] env[62952]: Removing descriptor: 19 [ 620.339802] env[62952]: ERROR nova.compute.manager [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1e4baadd-9333-4427-93c4-bad737fec2c6, please check neutron logs for more information. [ 620.339802] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Traceback (most recent call last): [ 620.339802] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 620.339802] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] yield resources [ 620.339802] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 620.339802] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] self.driver.spawn(context, instance, image_meta, [ 620.339802] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 620.339802] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] self._vmops.spawn(context, instance, image_meta, injected_files, [ 620.339802] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 620.339802] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] vm_ref = self.build_virtual_machine(instance, [ 620.339802] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 620.340486] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] vif_infos = vmwarevif.get_vif_info(self._session, [ 620.340486] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 620.340486] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] for vif in network_info: [ 620.340486] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 620.340486] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] return self._sync_wrapper(fn, *args, **kwargs) [ 620.340486] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 620.340486] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] self.wait() [ 620.340486] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 620.340486] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] self[:] = self._gt.wait() [ 620.340486] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 620.340486] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] return self._exit_event.wait() [ 620.340486] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 620.340486] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] result = hub.switch() [ 620.341232] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 620.341232] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] return self.greenlet.switch() [ 620.341232] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.341232] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] result = function(*args, **kwargs) [ 620.341232] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 620.341232] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] return func(*args, **kwargs) [ 620.341232] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 620.341232] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] raise e [ 620.341232] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.341232] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] nwinfo = self.network_api.allocate_for_instance( [ 620.341232] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 620.341232] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] created_port_ids = self._update_ports_for_instance( [ 620.341232] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 620.342536] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] with excutils.save_and_reraise_exception(): [ 620.342536] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.342536] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] self.force_reraise() [ 620.342536] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.342536] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] raise self.value [ 620.342536] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 620.342536] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] updated_port = self._update_port( [ 620.342536] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.342536] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] _ensure_no_port_binding_failure(port) [ 620.342536] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.342536] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] raise exception.PortBindingFailed(port_id=port['id']) [ 620.342536] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] nova.exception.PortBindingFailed: Binding failed for port 1e4baadd-9333-4427-93c4-bad737fec2c6, please check neutron logs for more information. [ 620.342536] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] [ 620.343104] env[62952]: INFO nova.compute.manager [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Terminating instance [ 620.346425] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Acquiring lock "refresh_cache-41dbbdcf-667c-4969-88da-43e42840ed20" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.346654] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Acquired lock "refresh_cache-41dbbdcf-667c-4969-88da-43e42840ed20" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.346903] env[62952]: DEBUG nova.network.neutron [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 620.643894] env[62952]: DEBUG oslo_concurrency.lockutils [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] Releasing lock "refresh_cache-dea74836-bdbb-4260-b15d-2dab04ee3940" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.644219] env[62952]: DEBUG nova.compute.manager [req-4fb43426-fa15-4d64-adc9-0ecd17f24212 req-299e86c2-a1bc-451c-9620-caaa46210a0c service nova] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Received event network-vif-deleted-b18d52c0-132e-44f5-a200-b977563b8c54 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 620.660537] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Acquiring lock "4c6f81ba-85f5-4ad8-a3c8-57867c613baa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.660779] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Lock "4c6f81ba-85f5-4ad8-a3c8-57867c613baa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.700618] env[62952]: DEBUG oslo_concurrency.lockutils [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "refresh_cache-9c762648-73d9-4976-90e2-13d6a394e7f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.700618] env[62952]: DEBUG nova.compute.manager [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 620.700618] env[62952]: DEBUG nova.compute.manager [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.700618] env[62952]: DEBUG nova.network.neutron [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 620.718311] env[62952]: DEBUG nova.network.neutron [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.793112] env[62952]: DEBUG nova.scheduler.client.report [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.901829] env[62952]: DEBUG nova.network.neutron [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.081792] env[62952]: DEBUG nova.network.neutron [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.224059] env[62952]: DEBUG nova.network.neutron [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.300570] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.926s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.301236] env[62952]: ERROR nova.compute.manager [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f2e7e01f-0ace-4b09-a54a-cbc7a33c5f3f, please check neutron logs for more information. [ 621.301236] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Traceback (most recent call last): [ 621.301236] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.301236] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] self.driver.spawn(context, instance, image_meta, [ 621.301236] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 621.301236] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.301236] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.301236] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] vm_ref = self.build_virtual_machine(instance, [ 621.301236] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.301236] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.301236] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.301604] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] for vif in network_info: [ 621.301604] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 621.301604] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] return self._sync_wrapper(fn, *args, **kwargs) [ 621.301604] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 621.301604] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] self.wait() [ 621.301604] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 621.301604] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] self[:] = self._gt.wait() [ 621.301604] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.301604] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] return self._exit_event.wait() [ 621.301604] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 621.301604] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] result = hub.switch() [ 621.301604] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 621.301604] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] return self.greenlet.switch() [ 621.301938] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.301938] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] result = function(*args, **kwargs) [ 621.301938] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 621.301938] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] return func(*args, **kwargs) [ 621.301938] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.301938] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] raise e [ 621.301938] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.301938] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] nwinfo = self.network_api.allocate_for_instance( [ 621.301938] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 621.301938] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] created_port_ids = self._update_ports_for_instance( [ 621.301938] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 621.301938] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] with excutils.save_and_reraise_exception(): [ 621.301938] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.302290] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] self.force_reraise() [ 621.302290] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.302290] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] raise self.value [ 621.302290] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 621.302290] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] updated_port = self._update_port( [ 621.302290] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.302290] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] _ensure_no_port_binding_failure(port) [ 621.302290] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.302290] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] raise exception.PortBindingFailed(port_id=port['id']) [ 621.302290] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] nova.exception.PortBindingFailed: Binding failed for port f2e7e01f-0ace-4b09-a54a-cbc7a33c5f3f, please check neutron logs for more information. [ 621.302290] env[62952]: ERROR nova.compute.manager [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] [ 621.302923] env[62952]: DEBUG nova.compute.utils [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Binding failed for port f2e7e01f-0ace-4b09-a54a-cbc7a33c5f3f, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 621.306170] env[62952]: DEBUG nova.compute.manager [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Build of instance 20cad4d8-5a23-47be-8603-2bf576700a14 was re-scheduled: Binding failed for port f2e7e01f-0ace-4b09-a54a-cbc7a33c5f3f, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 621.306170] env[62952]: DEBUG nova.compute.manager [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 621.306170] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Acquiring lock "refresh_cache-20cad4d8-5a23-47be-8603-2bf576700a14" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.306170] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Acquired lock "refresh_cache-20cad4d8-5a23-47be-8603-2bf576700a14" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.306388] env[62952]: DEBUG nova.network.neutron [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 621.306446] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.022s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.560575] env[62952]: DEBUG oslo_concurrency.lockutils [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquiring lock "f075161d-7447-4601-a06c-514fc4dd2eb9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.560929] env[62952]: DEBUG oslo_concurrency.lockutils [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Lock "f075161d-7447-4601-a06c-514fc4dd2eb9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.576561] env[62952]: DEBUG nova.compute.manager [req-86d97ff2-b26a-40e5-a38b-f6f6523bc9cc req-e89288b6-b1b3-4d79-a95f-a4ce0b205853 service nova] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Received event network-changed-1e4baadd-9333-4427-93c4-bad737fec2c6 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 621.576561] env[62952]: DEBUG nova.compute.manager [req-86d97ff2-b26a-40e5-a38b-f6f6523bc9cc req-e89288b6-b1b3-4d79-a95f-a4ce0b205853 service nova] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Refreshing instance network info cache due to event network-changed-1e4baadd-9333-4427-93c4-bad737fec2c6. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 621.576839] env[62952]: DEBUG oslo_concurrency.lockutils [req-86d97ff2-b26a-40e5-a38b-f6f6523bc9cc req-e89288b6-b1b3-4d79-a95f-a4ce0b205853 service nova] Acquiring lock "refresh_cache-41dbbdcf-667c-4969-88da-43e42840ed20" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.585280] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Releasing lock "refresh_cache-41dbbdcf-667c-4969-88da-43e42840ed20" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.585705] env[62952]: DEBUG nova.compute.manager [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 621.585880] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 621.586224] env[62952]: DEBUG oslo_concurrency.lockutils [req-86d97ff2-b26a-40e5-a38b-f6f6523bc9cc req-e89288b6-b1b3-4d79-a95f-a4ce0b205853 service nova] Acquired lock "refresh_cache-41dbbdcf-667c-4969-88da-43e42840ed20" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.586400] env[62952]: DEBUG nova.network.neutron [req-86d97ff2-b26a-40e5-a38b-f6f6523bc9cc req-e89288b6-b1b3-4d79-a95f-a4ce0b205853 service nova] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Refreshing network info cache for port 1e4baadd-9333-4427-93c4-bad737fec2c6 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 621.588338] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7bf0fc16-0d6f-4a19-8dd8-9c93253ac8e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.605299] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2257f2e-5758-4df4-9c83-318241c51778 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.628168] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 41dbbdcf-667c-4969-88da-43e42840ed20 could not be found. [ 621.628457] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 621.628674] env[62952]: INFO nova.compute.manager [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Took 0.04 seconds to destroy the instance on the hypervisor. [ 621.629069] env[62952]: DEBUG oslo.service.loopingcall [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 621.630205] env[62952]: DEBUG nova.compute.manager [-] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 621.630205] env[62952]: DEBUG nova.network.neutron [-] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 621.669507] env[62952]: DEBUG nova.network.neutron [-] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.726995] env[62952]: INFO nova.compute.manager [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] Took 1.03 seconds to deallocate network for instance. [ 621.842232] env[62952]: DEBUG nova.network.neutron [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.993499] env[62952]: DEBUG nova.network.neutron [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.015246] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Acquiring lock "d3559857-e983-4790-8283-05aa02437723" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.015474] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Lock "d3559857-e983-4790-8283-05aa02437723" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.098492] env[62952]: DEBUG oslo_concurrency.lockutils [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Acquiring lock "80a30ea8-e260-4b92-9099-be374333896f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.099363] env[62952]: DEBUG oslo_concurrency.lockutils [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Lock "80a30ea8-e260-4b92-9099-be374333896f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.129689] env[62952]: DEBUG nova.network.neutron [req-86d97ff2-b26a-40e5-a38b-f6f6523bc9cc req-e89288b6-b1b3-4d79-a95f-a4ce0b205853 service nova] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 622.173744] env[62952]: DEBUG nova.network.neutron [-] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.194199] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c65dcaa-132e-466f-b3fd-be773fe97b24 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.203094] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31601e72-ba32-42bb-8ee8-13cd8df4712f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.248729] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd08e609-fe91-47bc-afd0-58a1dc214bdf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.256343] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2723c15e-7591-4163-9865-f3e08c1382c8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.270914] env[62952]: DEBUG nova.compute.provider_tree [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.336659] env[62952]: DEBUG nova.network.neutron [req-86d97ff2-b26a-40e5-a38b-f6f6523bc9cc req-e89288b6-b1b3-4d79-a95f-a4ce0b205853 service nova] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.422449] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "3516272a-9ebd-413e-9b88-1a5c83b50f61" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.422673] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "3516272a-9ebd-413e-9b88-1a5c83b50f61" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.496321] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Releasing lock "refresh_cache-20cad4d8-5a23-47be-8603-2bf576700a14" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.496573] env[62952]: DEBUG nova.compute.manager [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 622.496739] env[62952]: DEBUG nova.compute.manager [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.496913] env[62952]: DEBUG nova.network.neutron [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 622.534553] env[62952]: DEBUG nova.network.neutron [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 622.680203] env[62952]: INFO nova.compute.manager [-] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Took 1.05 seconds to deallocate network for instance. [ 622.685341] env[62952]: DEBUG nova.compute.claims [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 622.685520] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.777248] env[62952]: DEBUG nova.scheduler.client.report [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 622.792455] env[62952]: INFO nova.scheduler.client.report [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Deleted allocations for instance 9c762648-73d9-4976-90e2-13d6a394e7f6 [ 622.839859] env[62952]: DEBUG oslo_concurrency.lockutils [req-86d97ff2-b26a-40e5-a38b-f6f6523bc9cc req-e89288b6-b1b3-4d79-a95f-a4ce0b205853 service nova] Releasing lock "refresh_cache-41dbbdcf-667c-4969-88da-43e42840ed20" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.038789] env[62952]: DEBUG nova.network.neutron [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.285213] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.978s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.288462] env[62952]: ERROR nova.compute.manager [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0c947d5e-7083-46d4-9b46-6e4a7331b432, please check neutron logs for more information. [ 623.288462] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Traceback (most recent call last): [ 623.288462] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 623.288462] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] self.driver.spawn(context, instance, image_meta, [ 623.288462] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 623.288462] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.288462] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.288462] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] vm_ref = self.build_virtual_machine(instance, [ 623.288462] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.288462] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.288462] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.288872] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] for vif in network_info: [ 623.288872] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 623.288872] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] return self._sync_wrapper(fn, *args, **kwargs) [ 623.288872] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 623.288872] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] self.wait() [ 623.288872] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 623.288872] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] self[:] = self._gt.wait() [ 623.288872] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.288872] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] return self._exit_event.wait() [ 623.288872] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.288872] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] result = hub.switch() [ 623.288872] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.288872] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] return self.greenlet.switch() [ 623.289234] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.289234] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] result = function(*args, **kwargs) [ 623.289234] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 623.289234] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] return func(*args, **kwargs) [ 623.289234] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.289234] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] raise e [ 623.289234] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.289234] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] nwinfo = self.network_api.allocate_for_instance( [ 623.289234] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 623.289234] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] created_port_ids = self._update_ports_for_instance( [ 623.289234] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 623.289234] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] with excutils.save_and_reraise_exception(): [ 623.289234] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.289628] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] self.force_reraise() [ 623.289628] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.289628] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] raise self.value [ 623.289628] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 623.289628] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] updated_port = self._update_port( [ 623.289628] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.289628] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] _ensure_no_port_binding_failure(port) [ 623.289628] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.289628] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] raise exception.PortBindingFailed(port_id=port['id']) [ 623.289628] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] nova.exception.PortBindingFailed: Binding failed for port 0c947d5e-7083-46d4-9b46-6e4a7331b432, please check neutron logs for more information. [ 623.289628] env[62952]: ERROR nova.compute.manager [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] [ 623.290226] env[62952]: DEBUG nova.compute.utils [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Binding failed for port 0c947d5e-7083-46d4-9b46-6e4a7331b432, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 623.290614] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.201s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.293322] env[62952]: DEBUG nova.compute.manager [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Build of instance 0e49e1b2-076a-4770-bbdc-3d5f7cf67590 was re-scheduled: Binding failed for port 0c947d5e-7083-46d4-9b46-6e4a7331b432, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 623.293862] env[62952]: DEBUG nova.compute.manager [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 623.293997] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Acquiring lock "refresh_cache-0e49e1b2-076a-4770-bbdc-3d5f7cf67590" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.295046] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Acquired lock "refresh_cache-0e49e1b2-076a-4770-bbdc-3d5f7cf67590" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.296252] env[62952]: DEBUG nova.network.neutron [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 623.301704] env[62952]: DEBUG oslo_concurrency.lockutils [None req-61033802-005c-45db-b84d-879b09368235 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "9c762648-73d9-4976-90e2-13d6a394e7f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.553s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.302880] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "9c762648-73d9-4976-90e2-13d6a394e7f6" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 19.430s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.303016] env[62952]: INFO nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 9c762648-73d9-4976-90e2-13d6a394e7f6] During sync_power_state the instance has a pending task (spawning). Skip. [ 623.303167] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "9c762648-73d9-4976-90e2-13d6a394e7f6" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.542860] env[62952]: INFO nova.compute.manager [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] Took 1.05 seconds to deallocate network for instance. [ 623.804344] env[62952]: DEBUG nova.compute.manager [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 623.869486] env[62952]: DEBUG nova.network.neutron [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.880551] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Acquiring lock "11139a80-26b2-4bc6-af91-18cd3adbe8fe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.880887] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Lock "11139a80-26b2-4bc6-af91-18cd3adbe8fe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.049119] env[62952]: DEBUG nova.network.neutron [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.123492] env[62952]: DEBUG nova.compute.manager [req-8f6ef448-2966-4f2b-9c2a-b75b96c593a0 req-4aa5dca0-8139-4ef9-b943-5693056b3d33 service nova] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Received event network-vif-deleted-1e4baadd-9333-4427-93c4-bad737fec2c6 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 624.171779] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-971a3894-2ed6-40ff-8c2e-bba9d2de8eef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.181492] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66366323-9051-4651-a613-e90bddeda692 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.217341] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb0c4bf9-7422-448b-90e3-53d812373a36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.225446] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1744949-a461-4e8b-ab0f-dcc92ccb6cf6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.240702] env[62952]: DEBUG nova.compute.provider_tree [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.336376] env[62952]: DEBUG oslo_concurrency.lockutils [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.558326] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Releasing lock "refresh_cache-0e49e1b2-076a-4770-bbdc-3d5f7cf67590" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.558326] env[62952]: DEBUG nova.compute.manager [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 624.558326] env[62952]: DEBUG nova.compute.manager [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 624.559766] env[62952]: DEBUG nova.network.neutron [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 624.592698] env[62952]: INFO nova.scheduler.client.report [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Deleted allocations for instance 20cad4d8-5a23-47be-8603-2bf576700a14 [ 624.602135] env[62952]: DEBUG nova.network.neutron [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.749198] env[62952]: DEBUG nova.scheduler.client.report [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.835892] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquiring lock "7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.836124] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.890373] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquiring lock "36f778a3-d795-467c-92d9-97e07b04956c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.890373] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "36f778a3-d795-467c-92d9-97e07b04956c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.949566] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquiring lock "7419de8d-5198-4152-a410-818af9c3b8e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.950069] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "7419de8d-5198-4152-a410-818af9c3b8e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.106682] env[62952]: DEBUG nova.network.neutron [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.109927] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b81420c7-0316-48ae-9a59-dac8d0ab2117 tempest-ServerDiagnosticsNegativeTest-1200630676 tempest-ServerDiagnosticsNegativeTest-1200630676-project-member] Lock "20cad4d8-5a23-47be-8603-2bf576700a14" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.474s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.109927] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "20cad4d8-5a23-47be-8603-2bf576700a14" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 21.237s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.111137] env[62952]: INFO nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 20cad4d8-5a23-47be-8603-2bf576700a14] During sync_power_state the instance has a pending task (spawning). Skip. [ 625.111496] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "20cad4d8-5a23-47be-8603-2bf576700a14" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.002s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.255692] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.965s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.256998] env[62952]: ERROR nova.compute.manager [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d5a318e4-4330-4101-a6d4-803e4a6f63ae, please check neutron logs for more information. [ 625.256998] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Traceback (most recent call last): [ 625.256998] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.256998] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] self.driver.spawn(context, instance, image_meta, [ 625.256998] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 625.256998] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.256998] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.256998] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] vm_ref = self.build_virtual_machine(instance, [ 625.256998] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.256998] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.256998] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.257966] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] for vif in network_info: [ 625.257966] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.257966] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] return self._sync_wrapper(fn, *args, **kwargs) [ 625.257966] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.257966] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] self.wait() [ 625.257966] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.257966] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] self[:] = self._gt.wait() [ 625.257966] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.257966] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] return self._exit_event.wait() [ 625.257966] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.257966] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] result = hub.switch() [ 625.257966] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.257966] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] return self.greenlet.switch() [ 625.259162] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.259162] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] result = function(*args, **kwargs) [ 625.259162] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.259162] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] return func(*args, **kwargs) [ 625.259162] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.259162] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] raise e [ 625.259162] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.259162] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] nwinfo = self.network_api.allocate_for_instance( [ 625.259162] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 625.259162] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] created_port_ids = self._update_ports_for_instance( [ 625.259162] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 625.259162] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] with excutils.save_and_reraise_exception(): [ 625.259162] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.259655] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] self.force_reraise() [ 625.259655] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.259655] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] raise self.value [ 625.259655] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 625.259655] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] updated_port = self._update_port( [ 625.259655] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.259655] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] _ensure_no_port_binding_failure(port) [ 625.259655] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.259655] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] raise exception.PortBindingFailed(port_id=port['id']) [ 625.259655] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] nova.exception.PortBindingFailed: Binding failed for port d5a318e4-4330-4101-a6d4-803e4a6f63ae, please check neutron logs for more information. [ 625.259655] env[62952]: ERROR nova.compute.manager [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] [ 625.259977] env[62952]: DEBUG nova.compute.utils [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Binding failed for port d5a318e4-4330-4101-a6d4-803e4a6f63ae, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 625.259977] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 11.957s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.262290] env[62952]: DEBUG nova.compute.manager [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Build of instance 4c40c507-b988-498c-b68a-a5304541a1c4 was re-scheduled: Binding failed for port d5a318e4-4330-4101-a6d4-803e4a6f63ae, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 625.263361] env[62952]: DEBUG nova.compute.manager [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 625.263464] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Acquiring lock "refresh_cache-4c40c507-b988-498c-b68a-a5304541a1c4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.263573] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Acquired lock "refresh_cache-4c40c507-b988-498c-b68a-a5304541a1c4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.264113] env[62952]: DEBUG nova.network.neutron [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 625.610530] env[62952]: INFO nova.compute.manager [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] Took 1.05 seconds to deallocate network for instance. [ 625.612998] env[62952]: DEBUG nova.compute.manager [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.808719] env[62952]: DEBUG nova.network.neutron [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.923214] env[62952]: DEBUG nova.network.neutron [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.150480] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.426829] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Releasing lock "refresh_cache-4c40c507-b988-498c-b68a-a5304541a1c4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.426974] env[62952]: DEBUG nova.compute.manager [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 626.428390] env[62952]: DEBUG nova.compute.manager [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 626.428390] env[62952]: DEBUG nova.network.neutron [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 626.455544] env[62952]: DEBUG nova.network.neutron [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.662192] env[62952]: INFO nova.scheduler.client.report [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Deleted allocations for instance 0e49e1b2-076a-4770-bbdc-3d5f7cf67590 [ 626.803192] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 0e49e1b2-076a-4770-bbdc-3d5f7cf67590 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 626.959631] env[62952]: DEBUG nova.network.neutron [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.174466] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba7cea86-f998-42ae-a81f-a78c1b22c071 tempest-ServerDiagnosticsTest-1626728970 tempest-ServerDiagnosticsTest-1626728970-project-member] Lock "0e49e1b2-076a-4770-bbdc-3d5f7cf67590" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.014s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.175619] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "0e49e1b2-076a-4770-bbdc-3d5f7cf67590" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 23.302s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.175862] env[62952]: INFO nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 0e49e1b2-076a-4770-bbdc-3d5f7cf67590] During sync_power_state the instance has a pending task (block_device_mapping). Skip. [ 627.176064] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "0e49e1b2-076a-4770-bbdc-3d5f7cf67590" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.311894] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 4c40c507-b988-498c-b68a-a5304541a1c4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 627.312504] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 1f51dd77-519f-4d66-9048-5f63e85e3837 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 627.312504] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 627.312504] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance dea74836-bdbb-4260-b15d-2dab04ee3940 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 627.312504] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 41dbbdcf-667c-4969-88da-43e42840ed20 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 627.465419] env[62952]: INFO nova.compute.manager [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] Took 1.04 seconds to deallocate network for instance. [ 627.681236] env[62952]: DEBUG nova.compute.manager [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 627.819020] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 6c87dd22-6c47-461a-8129-585d1cd72a2a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 628.102081] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Acquiring lock "0cdbad9a-9d52-45b0-8267-e181a5d38134" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.102081] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Lock "0cdbad9a-9d52-45b0-8267-e181a5d38134" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.211927] env[62952]: DEBUG oslo_concurrency.lockutils [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.320913] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance c377d16d-2641-483f-892b-42f6f067e9bc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 628.510581] env[62952]: INFO nova.scheduler.client.report [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Deleted allocations for instance 4c40c507-b988-498c-b68a-a5304541a1c4 [ 628.824897] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 25e10d02-8956-4d53-a1a4-b977da38f1c8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 629.019524] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17a9d000-3f54-48f1-928d-6f248137b032 tempest-TenantUsagesTestJSON-328215513 tempest-TenantUsagesTestJSON-328215513-project-member] Lock "4c40c507-b988-498c-b68a-a5304541a1c4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.723s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.020836] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "4c40c507-b988-498c-b68a-a5304541a1c4" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 25.147s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.021154] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-269a3c7f-8cb4-4c53-b7c5-b8641d98d8f3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.036155] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e006adaf-bf1a-47df-b908-ed5eb96852db {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.185978] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Acquiring lock "4bc891c1-e7e1-4196-9770-9bb42ed7499c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.186251] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Lock "4bc891c1-e7e1-4196-9770-9bb42ed7499c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.330351] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 78a0acf9-12f5-4699-a4b8-2cbefcbd343b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 629.531682] env[62952]: DEBUG nova.compute.manager [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 629.564942] env[62952]: INFO nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 4c40c507-b988-498c-b68a-a5304541a1c4] During the sync_power process the instance has moved from host None to host cpu-1 [ 629.565376] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "4c40c507-b988-498c-b68a-a5304541a1c4" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.544s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.834113] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 4eb9ea32-b2eb-440c-871b-6942590d9c01 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 630.071098] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.337774] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 630.844933] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 6ac19031-ac4e-4061-bbfd-75bd5c90803a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 631.349762] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance dc025711-45be-49b6-a765-f7687b94e6a1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 631.855649] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 0fd60040-c3ab-491f-a53e-f3297657367e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 632.358224] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 4c6f81ba-85f5-4ad8-a3c8-57867c613baa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 632.862294] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance f075161d-7447-4601-a06c-514fc4dd2eb9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 633.370878] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance d3559857-e983-4790-8283-05aa02437723 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 633.560213] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Acquiring lock "268e5249-0156-45dd-bbbd-10105d16f269" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.560213] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Lock "268e5249-0156-45dd-bbbd-10105d16f269" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.561471] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "f9ed97a4-ec32-415b-958e-d362b622b1d4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.561860] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "f9ed97a4-ec32-415b-958e-d362b622b1d4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.872382] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 80a30ea8-e260-4b92-9099-be374333896f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 634.376050] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 3516272a-9ebd-413e-9b88-1a5c83b50f61 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 634.880063] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 11139a80-26b2-4bc6-af91-18cd3adbe8fe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 635.384850] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 635.891499] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 36f778a3-d795-467c-92d9-97e07b04956c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 635.907629] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Acquiring lock "5922d142-0b6f-4479-abcb-0c067a1ef837" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.908304] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Lock "5922d142-0b6f-4479-abcb-0c067a1ef837" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.399317] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 7419de8d-5198-4152-a410-818af9c3b8e2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 636.399317] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 636.399317] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 636.825951] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29671bee-ccc4-4aee-a4b6-f0bbf1b94b32 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.833893] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23e8bede-2326-4761-b1ed-f35b1909f460 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.871819] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184e99f4-c2a0-4ea5-9961-c6dac72c6e6c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.879152] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6680275-9a65-4d09-8a1e-5868b2fd1e83 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.892083] env[62952]: DEBUG nova.compute.provider_tree [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.394980] env[62952]: DEBUG nova.scheduler.client.report [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.905027] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 637.905294] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 12.646s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.907697] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.754s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.909404] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 637.910025] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Getting list of instances from cluster (obj){ [ 637.910025] env[62952]: value = "domain-c8" [ 637.910025] env[62952]: _type = "ClusterComputeResource" [ 637.910025] env[62952]: } {{(pid=62952) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 637.911864] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df292d73-da1e-4bb0-aa1a-ea971a3ce51e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.920172] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Got total of 0 instances {{(pid=62952) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 638.591681] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Acquiring lock "a458a658-96f3-41b2-a9ee-2a6dd52bfeb4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.592261] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Lock "a458a658-96f3-41b2-a9ee-2a6dd52bfeb4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.871315] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdce4bd7-66be-468f-8f89-d679f0f8b203 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.881034] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d59fd84-aa1a-43f1-aa7f-96727fe8b617 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.921059] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451bfdda-e1bd-42a5-bf37-8daf3252da7b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.931172] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0946524-8437-48e8-a138-08ea2e1028df {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.946901] env[62952]: DEBUG nova.compute.provider_tree [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.450816] env[62952]: DEBUG nova.scheduler.client.report [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 639.958579] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.053s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.959248] env[62952]: ERROR nova.compute.manager [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2f040d58-ad7e-41c5-a6c7-27bc621f5b0d, please check neutron logs for more information. [ 639.959248] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Traceback (most recent call last): [ 639.959248] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 639.959248] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] self.driver.spawn(context, instance, image_meta, [ 639.959248] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 639.959248] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] self._vmops.spawn(context, instance, image_meta, injected_files, [ 639.959248] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 639.959248] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] vm_ref = self.build_virtual_machine(instance, [ 639.959248] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 639.959248] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] vif_infos = vmwarevif.get_vif_info(self._session, [ 639.959248] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 639.959668] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] for vif in network_info: [ 639.959668] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 639.959668] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] return self._sync_wrapper(fn, *args, **kwargs) [ 639.959668] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 639.959668] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] self.wait() [ 639.959668] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 639.959668] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] self[:] = self._gt.wait() [ 639.959668] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 639.959668] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] return self._exit_event.wait() [ 639.959668] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 639.959668] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] result = hub.switch() [ 639.959668] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 639.959668] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] return self.greenlet.switch() [ 639.960143] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.960143] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] result = function(*args, **kwargs) [ 639.960143] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 639.960143] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] return func(*args, **kwargs) [ 639.960143] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.960143] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] raise e [ 639.960143] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.960143] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] nwinfo = self.network_api.allocate_for_instance( [ 639.960143] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 639.960143] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] created_port_ids = self._update_ports_for_instance( [ 639.960143] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 639.960143] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] with excutils.save_and_reraise_exception(): [ 639.960143] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.960573] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] self.force_reraise() [ 639.960573] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.960573] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] raise self.value [ 639.960573] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 639.960573] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] updated_port = self._update_port( [ 639.960573] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.960573] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] _ensure_no_port_binding_failure(port) [ 639.960573] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.960573] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] raise exception.PortBindingFailed(port_id=port['id']) [ 639.960573] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] nova.exception.PortBindingFailed: Binding failed for port 2f040d58-ad7e-41c5-a6c7-27bc621f5b0d, please check neutron logs for more information. [ 639.960573] env[62952]: ERROR nova.compute.manager [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] [ 639.961082] env[62952]: DEBUG nova.compute.utils [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Binding failed for port 2f040d58-ad7e-41c5-a6c7-27bc621f5b0d, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 639.961835] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.660s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.968418] env[62952]: DEBUG nova.compute.manager [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Build of instance 1f51dd77-519f-4d66-9048-5f63e85e3837 was re-scheduled: Binding failed for port 2f040d58-ad7e-41c5-a6c7-27bc621f5b0d, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 639.968879] env[62952]: DEBUG nova.compute.manager [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 639.969673] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Acquiring lock "refresh_cache-1f51dd77-519f-4d66-9048-5f63e85e3837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.969673] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Acquired lock "refresh_cache-1f51dd77-519f-4d66-9048-5f63e85e3837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.969673] env[62952]: DEBUG nova.network.neutron [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 640.500411] env[62952]: DEBUG nova.network.neutron [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 640.609620] env[62952]: DEBUG nova.network.neutron [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.898067] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f367cb9-f800-446a-b0ad-83ed5237f506 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.905995] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b065521-4078-4f8c-a7d9-2f0b77f5d838 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.941800] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38fe49fe-6540-4a5a-b020-f26908edbfa8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.949975] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df8ed255-b8cb-462d-852f-91ec28be8204 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.963944] env[62952]: DEBUG nova.compute.provider_tree [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.114893] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Releasing lock "refresh_cache-1f51dd77-519f-4d66-9048-5f63e85e3837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.115019] env[62952]: DEBUG nova.compute.manager [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 641.115512] env[62952]: DEBUG nova.compute.manager [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 641.115512] env[62952]: DEBUG nova.network.neutron [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 641.134125] env[62952]: DEBUG nova.network.neutron [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.461037] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquiring lock "dd88d70d-17a3-4059-9092-5d6cec8b1ae4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.461508] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Lock "dd88d70d-17a3-4059-9092-5d6cec8b1ae4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.467647] env[62952]: DEBUG nova.scheduler.client.report [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 641.636609] env[62952]: DEBUG nova.network.neutron [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.976706] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.015s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.977525] env[62952]: ERROR nova.compute.manager [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 15b94fb4-8441-4c4d-bcbd-b444f31b2856, please check neutron logs for more information. [ 641.977525] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Traceback (most recent call last): [ 641.977525] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 641.977525] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] self.driver.spawn(context, instance, image_meta, [ 641.977525] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 641.977525] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.977525] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.977525] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] vm_ref = self.build_virtual_machine(instance, [ 641.977525] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.977525] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.977525] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.977930] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] for vif in network_info: [ 641.977930] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.977930] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] return self._sync_wrapper(fn, *args, **kwargs) [ 641.977930] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.977930] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] self.wait() [ 641.977930] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.977930] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] self[:] = self._gt.wait() [ 641.977930] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.977930] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] return self._exit_event.wait() [ 641.977930] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 641.977930] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] result = hub.switch() [ 641.977930] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 641.977930] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] return self.greenlet.switch() [ 641.978311] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.978311] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] result = function(*args, **kwargs) [ 641.978311] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.978311] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] return func(*args, **kwargs) [ 641.978311] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.978311] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] raise e [ 641.978311] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.978311] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] nwinfo = self.network_api.allocate_for_instance( [ 641.978311] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 641.978311] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] created_port_ids = self._update_ports_for_instance( [ 641.978311] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 641.978311] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] with excutils.save_and_reraise_exception(): [ 641.978311] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.978665] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] self.force_reraise() [ 641.978665] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.978665] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] raise self.value [ 641.978665] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 641.978665] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] updated_port = self._update_port( [ 641.978665] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.978665] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] _ensure_no_port_binding_failure(port) [ 641.978665] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.978665] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] raise exception.PortBindingFailed(port_id=port['id']) [ 641.978665] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] nova.exception.PortBindingFailed: Binding failed for port 15b94fb4-8441-4c4d-bcbd-b444f31b2856, please check neutron logs for more information. [ 641.978665] env[62952]: ERROR nova.compute.manager [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] [ 641.978964] env[62952]: DEBUG nova.compute.utils [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Binding failed for port 15b94fb4-8441-4c4d-bcbd-b444f31b2856, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 641.980543] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.306s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.982707] env[62952]: DEBUG nova.compute.manager [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Build of instance 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4 was re-scheduled: Binding failed for port 15b94fb4-8441-4c4d-bcbd-b444f31b2856, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 641.983013] env[62952]: DEBUG nova.compute.manager [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 641.983098] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Acquiring lock "refresh_cache-5db9f2cd-faf1-4972-8fa0-7b7705f40bf4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.983244] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Acquired lock "refresh_cache-5db9f2cd-faf1-4972-8fa0-7b7705f40bf4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.983406] env[62952]: DEBUG nova.network.neutron [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 642.139954] env[62952]: INFO nova.compute.manager [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] [instance: 1f51dd77-519f-4d66-9048-5f63e85e3837] Took 1.02 seconds to deallocate network for instance. [ 642.511546] env[62952]: DEBUG nova.network.neutron [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.569120] env[62952]: DEBUG nova.network.neutron [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.959568] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca58c563-b464-40f1-8493-163dfb13b706 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.964786] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6dc9f7b-64bc-4239-9269-6f217d4aa8a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.003046] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5700221f-281e-4494-bede-5bb8d79904e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.014558] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67088ba-336d-483e-8a45-3a695588ff1e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.028451] env[62952]: DEBUG nova.compute.provider_tree [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.075944] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Releasing lock "refresh_cache-5db9f2cd-faf1-4972-8fa0-7b7705f40bf4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.076296] env[62952]: DEBUG nova.compute.manager [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 643.076394] env[62952]: DEBUG nova.compute.manager [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.076560] env[62952]: DEBUG nova.network.neutron [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 643.093126] env[62952]: DEBUG nova.network.neutron [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.173135] env[62952]: INFO nova.scheduler.client.report [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Deleted allocations for instance 1f51dd77-519f-4d66-9048-5f63e85e3837 [ 643.534036] env[62952]: DEBUG nova.scheduler.client.report [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 643.596762] env[62952]: DEBUG nova.network.neutron [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.687450] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e79c9562-4928-4264-a2f1-4e2fed464776 tempest-FloatingIPsAssociationTestJSON-1947049472 tempest-FloatingIPsAssociationTestJSON-1947049472-project-member] Lock "1f51dd77-519f-4d66-9048-5f63e85e3837" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.021s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.036332] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.057s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.037339] env[62952]: ERROR nova.compute.manager [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b18d52c0-132e-44f5-a200-b977563b8c54, please check neutron logs for more information. [ 644.037339] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Traceback (most recent call last): [ 644.037339] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 644.037339] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] self.driver.spawn(context, instance, image_meta, [ 644.037339] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 644.037339] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] self._vmops.spawn(context, instance, image_meta, injected_files, [ 644.037339] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 644.037339] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] vm_ref = self.build_virtual_machine(instance, [ 644.037339] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 644.037339] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] vif_infos = vmwarevif.get_vif_info(self._session, [ 644.037339] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 644.037696] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] for vif in network_info: [ 644.037696] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 644.037696] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] return self._sync_wrapper(fn, *args, **kwargs) [ 644.037696] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 644.037696] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] self.wait() [ 644.037696] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 644.037696] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] self[:] = self._gt.wait() [ 644.037696] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 644.037696] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] return self._exit_event.wait() [ 644.037696] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 644.037696] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] result = hub.switch() [ 644.037696] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 644.037696] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] return self.greenlet.switch() [ 644.038108] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.038108] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] result = function(*args, **kwargs) [ 644.038108] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 644.038108] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] return func(*args, **kwargs) [ 644.038108] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.038108] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] raise e [ 644.038108] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.038108] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] nwinfo = self.network_api.allocate_for_instance( [ 644.038108] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.038108] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] created_port_ids = self._update_ports_for_instance( [ 644.038108] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.038108] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] with excutils.save_and_reraise_exception(): [ 644.038108] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.038471] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] self.force_reraise() [ 644.038471] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.038471] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] raise self.value [ 644.038471] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.038471] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] updated_port = self._update_port( [ 644.038471] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.038471] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] _ensure_no_port_binding_failure(port) [ 644.038471] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.038471] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] raise exception.PortBindingFailed(port_id=port['id']) [ 644.038471] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] nova.exception.PortBindingFailed: Binding failed for port b18d52c0-132e-44f5-a200-b977563b8c54, please check neutron logs for more information. [ 644.038471] env[62952]: ERROR nova.compute.manager [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] [ 644.040618] env[62952]: DEBUG nova.compute.utils [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Binding failed for port b18d52c0-132e-44f5-a200-b977563b8c54, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 644.041913] env[62952]: DEBUG oslo_concurrency.lockutils [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.829s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.043378] env[62952]: INFO nova.compute.claims [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 644.048883] env[62952]: DEBUG nova.compute.manager [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Build of instance dea74836-bdbb-4260-b15d-2dab04ee3940 was re-scheduled: Binding failed for port b18d52c0-132e-44f5-a200-b977563b8c54, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 644.048883] env[62952]: DEBUG nova.compute.manager [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 644.049265] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Acquiring lock "refresh_cache-dea74836-bdbb-4260-b15d-2dab04ee3940" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.049449] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Acquired lock "refresh_cache-dea74836-bdbb-4260-b15d-2dab04ee3940" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.049783] env[62952]: DEBUG nova.network.neutron [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 644.099052] env[62952]: INFO nova.compute.manager [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] [instance: 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4] Took 1.02 seconds to deallocate network for instance. [ 644.189477] env[62952]: DEBUG nova.compute.manager [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 644.571835] env[62952]: DEBUG nova.network.neutron [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.665646] env[62952]: DEBUG nova.network.neutron [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.721012] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.139913] env[62952]: INFO nova.scheduler.client.report [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Deleted allocations for instance 5db9f2cd-faf1-4972-8fa0-7b7705f40bf4 [ 645.169117] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Releasing lock "refresh_cache-dea74836-bdbb-4260-b15d-2dab04ee3940" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.169117] env[62952]: DEBUG nova.compute.manager [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 645.169117] env[62952]: DEBUG nova.compute.manager [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 645.169413] env[62952]: DEBUG nova.network.neutron [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 645.190315] env[62952]: DEBUG nova.network.neutron [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.485347] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d81442-d8f0-4ee7-91c0-1a7ea5e7bfb4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.493284] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283baec9-a40a-45f7-af2c-764ce6b2053f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.527618] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b124f2c-c19e-4f5f-9c5c-0f340fa05747 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.537301] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdecae0b-db01-4e08-9109-6c4586ff948e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.551938] env[62952]: DEBUG nova.compute.provider_tree [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.651717] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a35047f5-e98c-4d29-9583-2b449f4e64ac tempest-ServerExternalEventsTest-920452419 tempest-ServerExternalEventsTest-920452419-project-member] Lock "5db9f2cd-faf1-4972-8fa0-7b7705f40bf4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.189s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.693072] env[62952]: DEBUG nova.network.neutron [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.058050] env[62952]: DEBUG nova.scheduler.client.report [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 646.155489] env[62952]: DEBUG nova.compute.manager [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 646.195897] env[62952]: INFO nova.compute.manager [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: dea74836-bdbb-4260-b15d-2dab04ee3940] Took 1.03 seconds to deallocate network for instance. [ 646.563823] env[62952]: DEBUG oslo_concurrency.lockutils [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.520s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.563823] env[62952]: DEBUG nova.compute.manager [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 646.569089] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.461s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.572160] env[62952]: INFO nova.compute.claims [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 646.691377] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.070938] env[62952]: DEBUG nova.compute.utils [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 647.072405] env[62952]: DEBUG nova.compute.manager [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 647.072607] env[62952]: DEBUG nova.network.neutron [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 647.136824] env[62952]: DEBUG nova.policy [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ff3a53dbede4951b062ac5df0d1c011', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '02ecd052199f4ec4a291755df3cd4d2e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 647.242760] env[62952]: INFO nova.scheduler.client.report [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Deleted allocations for instance dea74836-bdbb-4260-b15d-2dab04ee3940 [ 647.580206] env[62952]: DEBUG nova.compute.manager [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 647.659245] env[62952]: DEBUG nova.network.neutron [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Successfully created port: a5b1d0ee-1acc-4231-9031-c18558472a71 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 647.760667] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af7498b4-16ed-4e34-b9d4-ce30c3ad7cc3 tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Lock "dea74836-bdbb-4260-b15d-2dab04ee3940" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.770s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.018194] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ca88a6-e6ab-475d-9422-616e9ece3c68 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.027630] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd4f923-8a20-4110-957a-d5b66ccbcc27 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.060758] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71895dc1-db58-4a46-9608-a86f4ccd253a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.069872] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51da2d22-9069-4726-ad13-2ad42984617b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.094936] env[62952]: DEBUG nova.compute.provider_tree [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.263513] env[62952]: DEBUG nova.compute.manager [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 648.597041] env[62952]: DEBUG nova.compute.manager [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 648.601087] env[62952]: DEBUG nova.scheduler.client.report [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 648.641179] env[62952]: DEBUG nova.virt.hardware [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 648.641441] env[62952]: DEBUG nova.virt.hardware [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 648.641621] env[62952]: DEBUG nova.virt.hardware [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 648.641810] env[62952]: DEBUG nova.virt.hardware [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 648.642059] env[62952]: DEBUG nova.virt.hardware [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 648.642140] env[62952]: DEBUG nova.virt.hardware [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 648.642343] env[62952]: DEBUG nova.virt.hardware [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 648.642504] env[62952]: DEBUG nova.virt.hardware [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 648.642647] env[62952]: DEBUG nova.virt.hardware [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 648.642820] env[62952]: DEBUG nova.virt.hardware [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 648.643008] env[62952]: DEBUG nova.virt.hardware [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 648.644432] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-434bd322-0eac-4ada-8e2f-cdcfbfa42691 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.655373] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2f30aa6-f852-432b-93a1-87fd706d48aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.789657] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.050217] env[62952]: DEBUG nova.compute.manager [req-c540848d-065e-4af0-8060-a94a575ff309 req-31ca9985-b139-43c0-9494-700bc45bdc28 service nova] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Received event network-changed-a5b1d0ee-1acc-4231-9031-c18558472a71 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 649.050595] env[62952]: DEBUG nova.compute.manager [req-c540848d-065e-4af0-8060-a94a575ff309 req-31ca9985-b139-43c0-9494-700bc45bdc28 service nova] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Refreshing instance network info cache due to event network-changed-a5b1d0ee-1acc-4231-9031-c18558472a71. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 649.051104] env[62952]: DEBUG oslo_concurrency.lockutils [req-c540848d-065e-4af0-8060-a94a575ff309 req-31ca9985-b139-43c0-9494-700bc45bdc28 service nova] Acquiring lock "refresh_cache-6c87dd22-6c47-461a-8129-585d1cd72a2a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.051283] env[62952]: DEBUG oslo_concurrency.lockutils [req-c540848d-065e-4af0-8060-a94a575ff309 req-31ca9985-b139-43c0-9494-700bc45bdc28 service nova] Acquired lock "refresh_cache-6c87dd22-6c47-461a-8129-585d1cd72a2a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.051498] env[62952]: DEBUG nova.network.neutron [req-c540848d-065e-4af0-8060-a94a575ff309 req-31ca9985-b139-43c0-9494-700bc45bdc28 service nova] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Refreshing network info cache for port a5b1d0ee-1acc-4231-9031-c18558472a71 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 649.107880] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.539s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.108758] env[62952]: DEBUG nova.compute.manager [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 649.116127] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.429s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.221052] env[62952]: ERROR nova.compute.manager [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a5b1d0ee-1acc-4231-9031-c18558472a71, please check neutron logs for more information. [ 649.221052] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 649.221052] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.221052] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 649.221052] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 649.221052] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 649.221052] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 649.221052] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 649.221052] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.221052] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 649.221052] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.221052] env[62952]: ERROR nova.compute.manager raise self.value [ 649.221052] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 649.221052] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 649.221052] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.221052] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 649.221659] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.221659] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 649.221659] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a5b1d0ee-1acc-4231-9031-c18558472a71, please check neutron logs for more information. [ 649.221659] env[62952]: ERROR nova.compute.manager [ 649.221659] env[62952]: Traceback (most recent call last): [ 649.221659] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 649.221659] env[62952]: listener.cb(fileno) [ 649.221659] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.221659] env[62952]: result = function(*args, **kwargs) [ 649.221659] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 649.221659] env[62952]: return func(*args, **kwargs) [ 649.221659] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 649.221659] env[62952]: raise e [ 649.221659] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.221659] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 649.221659] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 649.221659] env[62952]: created_port_ids = self._update_ports_for_instance( [ 649.221659] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 649.221659] env[62952]: with excutils.save_and_reraise_exception(): [ 649.221659] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.221659] env[62952]: self.force_reraise() [ 649.221659] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.221659] env[62952]: raise self.value [ 649.221659] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 649.221659] env[62952]: updated_port = self._update_port( [ 649.221659] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.221659] env[62952]: _ensure_no_port_binding_failure(port) [ 649.221659] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.221659] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 649.223036] env[62952]: nova.exception.PortBindingFailed: Binding failed for port a5b1d0ee-1acc-4231-9031-c18558472a71, please check neutron logs for more information. [ 649.223036] env[62952]: Removing descriptor: 19 [ 649.223036] env[62952]: ERROR nova.compute.manager [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a5b1d0ee-1acc-4231-9031-c18558472a71, please check neutron logs for more information. [ 649.223036] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Traceback (most recent call last): [ 649.223036] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 649.223036] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] yield resources [ 649.223036] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 649.223036] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] self.driver.spawn(context, instance, image_meta, [ 649.223036] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 649.223036] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 649.223036] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 649.223036] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] vm_ref = self.build_virtual_machine(instance, [ 649.223379] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 649.223379] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] vif_infos = vmwarevif.get_vif_info(self._session, [ 649.223379] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 649.223379] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] for vif in network_info: [ 649.223379] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 649.223379] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] return self._sync_wrapper(fn, *args, **kwargs) [ 649.223379] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 649.223379] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] self.wait() [ 649.223379] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 649.223379] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] self[:] = self._gt.wait() [ 649.223379] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 649.223379] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] return self._exit_event.wait() [ 649.223379] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 649.223747] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] result = hub.switch() [ 649.223747] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 649.223747] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] return self.greenlet.switch() [ 649.223747] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.223747] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] result = function(*args, **kwargs) [ 649.223747] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 649.223747] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] return func(*args, **kwargs) [ 649.223747] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 649.223747] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] raise e [ 649.223747] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.223747] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] nwinfo = self.network_api.allocate_for_instance( [ 649.223747] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 649.223747] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] created_port_ids = self._update_ports_for_instance( [ 649.224189] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 649.224189] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] with excutils.save_and_reraise_exception(): [ 649.224189] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.224189] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] self.force_reraise() [ 649.224189] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.224189] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] raise self.value [ 649.224189] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 649.224189] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] updated_port = self._update_port( [ 649.224189] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.224189] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] _ensure_no_port_binding_failure(port) [ 649.224189] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.224189] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] raise exception.PortBindingFailed(port_id=port['id']) [ 649.224560] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] nova.exception.PortBindingFailed: Binding failed for port a5b1d0ee-1acc-4231-9031-c18558472a71, please check neutron logs for more information. [ 649.224560] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] [ 649.224560] env[62952]: INFO nova.compute.manager [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Terminating instance [ 649.231139] env[62952]: DEBUG oslo_concurrency.lockutils [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Acquiring lock "refresh_cache-6c87dd22-6c47-461a-8129-585d1cd72a2a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.574299] env[62952]: DEBUG nova.network.neutron [req-c540848d-065e-4af0-8060-a94a575ff309 req-31ca9985-b139-43c0-9494-700bc45bdc28 service nova] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.624987] env[62952]: DEBUG nova.compute.utils [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 649.633312] env[62952]: DEBUG nova.compute.manager [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 649.633687] env[62952]: DEBUG nova.network.neutron [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 649.717093] env[62952]: DEBUG nova.network.neutron [req-c540848d-065e-4af0-8060-a94a575ff309 req-31ca9985-b139-43c0-9494-700bc45bdc28 service nova] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.742891] env[62952]: DEBUG nova.policy [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2c033769a01c4431bd1e1776c93d685e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ee5576448adf4151ac556f3d2f9aeeb3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 650.107755] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b9920d-3e43-4d99-8263-fc640e5e53b0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.121159] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a09749-24e4-47b1-9607-001cb9b7a44b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.155319] env[62952]: DEBUG nova.compute.manager [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 650.163734] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3402abf8-750a-4e13-9a5e-b683a2e71bb3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.172611] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e1b9aa-638a-48fc-877c-e00d15ebd148 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.187906] env[62952]: DEBUG nova.compute.provider_tree [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.220113] env[62952]: DEBUG oslo_concurrency.lockutils [req-c540848d-065e-4af0-8060-a94a575ff309 req-31ca9985-b139-43c0-9494-700bc45bdc28 service nova] Releasing lock "refresh_cache-6c87dd22-6c47-461a-8129-585d1cd72a2a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.220538] env[62952]: DEBUG oslo_concurrency.lockutils [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Acquired lock "refresh_cache-6c87dd22-6c47-461a-8129-585d1cd72a2a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.220689] env[62952]: DEBUG nova.network.neutron [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 650.360066] env[62952]: DEBUG nova.network.neutron [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Successfully created port: 0821cf2b-e08b-4f53-a4b6-da998fa92890 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 650.695020] env[62952]: DEBUG nova.scheduler.client.report [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 650.763172] env[62952]: DEBUG nova.network.neutron [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.866280] env[62952]: DEBUG nova.network.neutron [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.940340] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquiring lock "472e60f1-9b6c-415e-8a92-19235ad2e547" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.940340] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Lock "472e60f1-9b6c-415e-8a92-19235ad2e547" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.183191] env[62952]: DEBUG nova.compute.manager [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 651.204039] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.089s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.204537] env[62952]: ERROR nova.compute.manager [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1e4baadd-9333-4427-93c4-bad737fec2c6, please check neutron logs for more information. [ 651.204537] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Traceback (most recent call last): [ 651.204537] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 651.204537] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] self.driver.spawn(context, instance, image_meta, [ 651.204537] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 651.204537] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] self._vmops.spawn(context, instance, image_meta, injected_files, [ 651.204537] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 651.204537] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] vm_ref = self.build_virtual_machine(instance, [ 651.204537] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 651.204537] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] vif_infos = vmwarevif.get_vif_info(self._session, [ 651.204537] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 651.204964] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] for vif in network_info: [ 651.204964] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 651.204964] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] return self._sync_wrapper(fn, *args, **kwargs) [ 651.204964] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 651.204964] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] self.wait() [ 651.204964] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 651.204964] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] self[:] = self._gt.wait() [ 651.204964] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 651.204964] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] return self._exit_event.wait() [ 651.204964] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 651.204964] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] result = hub.switch() [ 651.204964] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 651.204964] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] return self.greenlet.switch() [ 651.205359] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.205359] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] result = function(*args, **kwargs) [ 651.205359] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 651.205359] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] return func(*args, **kwargs) [ 651.205359] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.205359] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] raise e [ 651.205359] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.205359] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] nwinfo = self.network_api.allocate_for_instance( [ 651.205359] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 651.205359] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] created_port_ids = self._update_ports_for_instance( [ 651.205359] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 651.205359] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] with excutils.save_and_reraise_exception(): [ 651.205359] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.205749] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] self.force_reraise() [ 651.205749] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.205749] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] raise self.value [ 651.205749] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 651.205749] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] updated_port = self._update_port( [ 651.205749] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.205749] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] _ensure_no_port_binding_failure(port) [ 651.205749] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.205749] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] raise exception.PortBindingFailed(port_id=port['id']) [ 651.205749] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] nova.exception.PortBindingFailed: Binding failed for port 1e4baadd-9333-4427-93c4-bad737fec2c6, please check neutron logs for more information. [ 651.205749] env[62952]: ERROR nova.compute.manager [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] [ 651.206088] env[62952]: DEBUG nova.compute.utils [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Binding failed for port 1e4baadd-9333-4427-93c4-bad737fec2c6, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 651.206401] env[62952]: DEBUG oslo_concurrency.lockutils [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.870s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.210881] env[62952]: INFO nova.compute.claims [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 651.217202] env[62952]: DEBUG nova.compute.manager [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Build of instance 41dbbdcf-667c-4969-88da-43e42840ed20 was re-scheduled: Binding failed for port 1e4baadd-9333-4427-93c4-bad737fec2c6, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 651.217202] env[62952]: DEBUG nova.compute.manager [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 651.217202] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Acquiring lock "refresh_cache-41dbbdcf-667c-4969-88da-43e42840ed20" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.217202] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Acquired lock "refresh_cache-41dbbdcf-667c-4969-88da-43e42840ed20" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.217472] env[62952]: DEBUG nova.network.neutron [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 651.235241] env[62952]: DEBUG nova.virt.hardware [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 651.235540] env[62952]: DEBUG nova.virt.hardware [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 651.235694] env[62952]: DEBUG nova.virt.hardware [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 651.236438] env[62952]: DEBUG nova.virt.hardware [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 651.236438] env[62952]: DEBUG nova.virt.hardware [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 651.236438] env[62952]: DEBUG nova.virt.hardware [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 651.236438] env[62952]: DEBUG nova.virt.hardware [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 651.236778] env[62952]: DEBUG nova.virt.hardware [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 651.236778] env[62952]: DEBUG nova.virt.hardware [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 651.236848] env[62952]: DEBUG nova.virt.hardware [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 651.237828] env[62952]: DEBUG nova.virt.hardware [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 651.240156] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852dd663-8e14-4733-a7a6-e371da4ecc39 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.252933] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace1bb9b-2c6d-489c-ba0b-250eb3f531aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.371448] env[62952]: DEBUG oslo_concurrency.lockutils [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Releasing lock "refresh_cache-6c87dd22-6c47-461a-8129-585d1cd72a2a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.372069] env[62952]: DEBUG nova.compute.manager [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 651.372174] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 651.372416] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c231173-4fbb-4209-9e1c-b7400d574f42 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.384868] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de9bb273-3046-4136-95d4-24515ebfabc3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.412185] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6c87dd22-6c47-461a-8129-585d1cd72a2a could not be found. [ 651.412418] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 651.412596] env[62952]: INFO nova.compute.manager [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 651.412955] env[62952]: DEBUG oslo.service.loopingcall [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 651.413473] env[62952]: DEBUG nova.compute.manager [-] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 651.413473] env[62952]: DEBUG nova.network.neutron [-] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 651.441245] env[62952]: DEBUG nova.network.neutron [-] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.756766] env[62952]: DEBUG nova.network.neutron [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.844889] env[62952]: DEBUG nova.compute.manager [req-07b87cc7-5527-4879-aee8-4253a35a374f req-43b87d34-34b9-41ea-9907-3d95ebc79264 service nova] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Received event network-vif-deleted-a5b1d0ee-1acc-4231-9031-c18558472a71 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 651.889820] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.890168] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.932382] env[62952]: DEBUG nova.network.neutron [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.945053] env[62952]: DEBUG nova.network.neutron [-] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.078426] env[62952]: ERROR nova.compute.manager [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0821cf2b-e08b-4f53-a4b6-da998fa92890, please check neutron logs for more information. [ 652.078426] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 652.078426] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.078426] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 652.078426] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.078426] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 652.078426] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.078426] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 652.078426] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.078426] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 652.078426] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.078426] env[62952]: ERROR nova.compute.manager raise self.value [ 652.078426] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.078426] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 652.078426] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.078426] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 652.078917] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.078917] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 652.078917] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0821cf2b-e08b-4f53-a4b6-da998fa92890, please check neutron logs for more information. [ 652.078917] env[62952]: ERROR nova.compute.manager [ 652.078917] env[62952]: Traceback (most recent call last): [ 652.078917] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 652.078917] env[62952]: listener.cb(fileno) [ 652.078917] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.078917] env[62952]: result = function(*args, **kwargs) [ 652.078917] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.078917] env[62952]: return func(*args, **kwargs) [ 652.078917] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.078917] env[62952]: raise e [ 652.078917] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.078917] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 652.078917] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.078917] env[62952]: created_port_ids = self._update_ports_for_instance( [ 652.078917] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.078917] env[62952]: with excutils.save_and_reraise_exception(): [ 652.078917] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.078917] env[62952]: self.force_reraise() [ 652.078917] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.078917] env[62952]: raise self.value [ 652.078917] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.078917] env[62952]: updated_port = self._update_port( [ 652.078917] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.078917] env[62952]: _ensure_no_port_binding_failure(port) [ 652.078917] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.078917] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 652.080202] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 0821cf2b-e08b-4f53-a4b6-da998fa92890, please check neutron logs for more information. [ 652.080202] env[62952]: Removing descriptor: 15 [ 652.080202] env[62952]: ERROR nova.compute.manager [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0821cf2b-e08b-4f53-a4b6-da998fa92890, please check neutron logs for more information. [ 652.080202] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Traceback (most recent call last): [ 652.080202] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 652.080202] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] yield resources [ 652.080202] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 652.080202] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] self.driver.spawn(context, instance, image_meta, [ 652.080202] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 652.080202] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.080202] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.080202] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] vm_ref = self.build_virtual_machine(instance, [ 652.080763] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.080763] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.080763] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.080763] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] for vif in network_info: [ 652.080763] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 652.080763] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] return self._sync_wrapper(fn, *args, **kwargs) [ 652.080763] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 652.080763] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] self.wait() [ 652.080763] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 652.080763] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] self[:] = self._gt.wait() [ 652.080763] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.080763] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] return self._exit_event.wait() [ 652.080763] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 652.081345] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] result = hub.switch() [ 652.081345] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 652.081345] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] return self.greenlet.switch() [ 652.081345] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.081345] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] result = function(*args, **kwargs) [ 652.081345] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.081345] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] return func(*args, **kwargs) [ 652.081345] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.081345] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] raise e [ 652.081345] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.081345] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] nwinfo = self.network_api.allocate_for_instance( [ 652.081345] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.081345] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] created_port_ids = self._update_ports_for_instance( [ 652.081762] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.081762] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] with excutils.save_and_reraise_exception(): [ 652.081762] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.081762] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] self.force_reraise() [ 652.081762] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.081762] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] raise self.value [ 652.081762] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.081762] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] updated_port = self._update_port( [ 652.081762] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.081762] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] _ensure_no_port_binding_failure(port) [ 652.081762] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.081762] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] raise exception.PortBindingFailed(port_id=port['id']) [ 652.082242] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] nova.exception.PortBindingFailed: Binding failed for port 0821cf2b-e08b-4f53-a4b6-da998fa92890, please check neutron logs for more information. [ 652.082242] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] [ 652.082242] env[62952]: INFO nova.compute.manager [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Terminating instance [ 652.082242] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Acquiring lock "refresh_cache-c377d16d-2641-483f-892b-42f6f067e9bc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.082438] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Acquired lock "refresh_cache-c377d16d-2641-483f-892b-42f6f067e9bc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.083343] env[62952]: DEBUG nova.network.neutron [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 652.436148] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Releasing lock "refresh_cache-41dbbdcf-667c-4969-88da-43e42840ed20" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.436464] env[62952]: DEBUG nova.compute.manager [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 652.436722] env[62952]: DEBUG nova.compute.manager [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 652.439087] env[62952]: DEBUG nova.network.neutron [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 652.451291] env[62952]: INFO nova.compute.manager [-] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Took 1.04 seconds to deallocate network for instance. [ 652.452998] env[62952]: DEBUG nova.compute.claims [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 652.453202] env[62952]: DEBUG oslo_concurrency.lockutils [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.454082] env[62952]: DEBUG nova.network.neutron [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 652.608499] env[62952]: DEBUG nova.network.neutron [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 652.732331] env[62952]: DEBUG nova.network.neutron [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.770265] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989c28a1-423b-4572-8e06-2ef3e831c74d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.782050] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-445123bb-e6b5-4e30-8dfa-28916d551427 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.816361] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4eb8425-9148-4bf8-ad34-9627c677aad0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.825392] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c8b6c6a-e802-4717-830d-58a8d439a1a7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.839832] env[62952]: DEBUG nova.compute.provider_tree [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.960420] env[62952]: DEBUG nova.network.neutron [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.241131] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Releasing lock "refresh_cache-c377d16d-2641-483f-892b-42f6f067e9bc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.241131] env[62952]: DEBUG nova.compute.manager [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 653.241131] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 653.241131] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ca99df1-1db4-4976-a0dc-6c3ab35c2932 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.250704] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d8fae4-7714-4f07-8e74-97a9e85020da {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.278739] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c377d16d-2641-483f-892b-42f6f067e9bc could not be found. [ 653.279144] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 653.279414] env[62952]: INFO nova.compute.manager [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Took 0.04 seconds to destroy the instance on the hypervisor. [ 653.279654] env[62952]: DEBUG oslo.service.loopingcall [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 653.279868] env[62952]: DEBUG nova.compute.manager [-] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.279959] env[62952]: DEBUG nova.network.neutron [-] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 653.302452] env[62952]: DEBUG nova.network.neutron [-] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 653.344617] env[62952]: DEBUG nova.scheduler.client.report [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 653.466124] env[62952]: INFO nova.compute.manager [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] [instance: 41dbbdcf-667c-4969-88da-43e42840ed20] Took 1.03 seconds to deallocate network for instance. [ 653.534176] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Acquiring lock "da96503a-8ae8-4b68-a782-631ee98754f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.534176] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Lock "da96503a-8ae8-4b68-a782-631ee98754f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.734127] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Acquiring lock "7f77fa00-27be-4a75-aac0-e1918f817837" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.734127] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Lock "7f77fa00-27be-4a75-aac0-e1918f817837" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.806575] env[62952]: DEBUG nova.network.neutron [-] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.854099] env[62952]: DEBUG oslo_concurrency.lockutils [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.645s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.854099] env[62952]: DEBUG nova.compute.manager [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 653.854942] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.705s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.862432] env[62952]: INFO nova.compute.claims [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 653.902572] env[62952]: DEBUG nova.compute.manager [req-f397d0ff-ed7d-41c8-9380-1014f41c7c11 req-0f0a2b9d-7f3c-40d1-aee8-702795a24f47 service nova] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Received event network-changed-0821cf2b-e08b-4f53-a4b6-da998fa92890 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 653.902772] env[62952]: DEBUG nova.compute.manager [req-f397d0ff-ed7d-41c8-9380-1014f41c7c11 req-0f0a2b9d-7f3c-40d1-aee8-702795a24f47 service nova] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Refreshing instance network info cache due to event network-changed-0821cf2b-e08b-4f53-a4b6-da998fa92890. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 653.902987] env[62952]: DEBUG oslo_concurrency.lockutils [req-f397d0ff-ed7d-41c8-9380-1014f41c7c11 req-0f0a2b9d-7f3c-40d1-aee8-702795a24f47 service nova] Acquiring lock "refresh_cache-c377d16d-2641-483f-892b-42f6f067e9bc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.903143] env[62952]: DEBUG oslo_concurrency.lockutils [req-f397d0ff-ed7d-41c8-9380-1014f41c7c11 req-0f0a2b9d-7f3c-40d1-aee8-702795a24f47 service nova] Acquired lock "refresh_cache-c377d16d-2641-483f-892b-42f6f067e9bc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.903300] env[62952]: DEBUG nova.network.neutron [req-f397d0ff-ed7d-41c8-9380-1014f41c7c11 req-0f0a2b9d-7f3c-40d1-aee8-702795a24f47 service nova] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Refreshing network info cache for port 0821cf2b-e08b-4f53-a4b6-da998fa92890 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 654.308685] env[62952]: INFO nova.compute.manager [-] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Took 1.03 seconds to deallocate network for instance. [ 654.311401] env[62952]: DEBUG nova.compute.claims [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 654.311583] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.367464] env[62952]: DEBUG nova.compute.utils [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 654.370974] env[62952]: DEBUG nova.compute.manager [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 654.370974] env[62952]: DEBUG nova.network.neutron [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 654.435320] env[62952]: DEBUG nova.network.neutron [req-f397d0ff-ed7d-41c8-9380-1014f41c7c11 req-0f0a2b9d-7f3c-40d1-aee8-702795a24f47 service nova] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.451640] env[62952]: DEBUG nova.policy [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ad058be856d6490d97ca0d02d0c341e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0bd4fcd3eb9744ad813c98deb07559a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 654.504654] env[62952]: INFO nova.scheduler.client.report [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Deleted allocations for instance 41dbbdcf-667c-4969-88da-43e42840ed20 [ 654.728399] env[62952]: DEBUG nova.network.neutron [req-f397d0ff-ed7d-41c8-9380-1014f41c7c11 req-0f0a2b9d-7f3c-40d1-aee8-702795a24f47 service nova] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.871572] env[62952]: DEBUG nova.compute.manager [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 654.973825] env[62952]: DEBUG nova.network.neutron [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Successfully created port: 270db801-02b2-419b-9df9-eb2c926b8e24 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 655.016910] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4334f9c7-fef0-47cf-b4a4-88d79c6806f2 tempest-ServersWithSpecificFlavorTestJSON-2085304920 tempest-ServersWithSpecificFlavorTestJSON-2085304920-project-member] Lock "41dbbdcf-667c-4969-88da-43e42840ed20" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.875s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.231750] env[62952]: DEBUG oslo_concurrency.lockutils [req-f397d0ff-ed7d-41c8-9380-1014f41c7c11 req-0f0a2b9d-7f3c-40d1-aee8-702795a24f47 service nova] Releasing lock "refresh_cache-c377d16d-2641-483f-892b-42f6f067e9bc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.231959] env[62952]: DEBUG nova.compute.manager [req-f397d0ff-ed7d-41c8-9380-1014f41c7c11 req-0f0a2b9d-7f3c-40d1-aee8-702795a24f47 service nova] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Received event network-vif-deleted-0821cf2b-e08b-4f53-a4b6-da998fa92890 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 655.302959] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df358510-e5f1-404d-9931-4ba07407ed42 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.313023] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8187c314-5fb4-4026-a7fc-4d8220889f7e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.350760] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-582e74e7-59d7-4b9e-b095-03b90b002b70 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.360498] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e4fa560-8e72-4834-b2ae-c33a7b3e05f8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.376420] env[62952]: DEBUG nova.compute.provider_tree [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.519913] env[62952]: DEBUG nova.compute.manager [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 655.656886] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Acquiring lock "879dc564-0a5e-48aa-92ff-b02435445a68" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.657745] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Lock "879dc564-0a5e-48aa-92ff-b02435445a68" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.882590] env[62952]: DEBUG nova.scheduler.client.report [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.886401] env[62952]: DEBUG nova.compute.manager [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 655.920170] env[62952]: DEBUG nova.virt.hardware [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 655.920170] env[62952]: DEBUG nova.virt.hardware [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 655.920170] env[62952]: DEBUG nova.virt.hardware [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 655.920332] env[62952]: DEBUG nova.virt.hardware [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 655.920332] env[62952]: DEBUG nova.virt.hardware [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 655.920332] env[62952]: DEBUG nova.virt.hardware [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 655.920332] env[62952]: DEBUG nova.virt.hardware [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 655.920332] env[62952]: DEBUG nova.virt.hardware [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 655.920512] env[62952]: DEBUG nova.virt.hardware [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 655.920512] env[62952]: DEBUG nova.virt.hardware [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 655.920512] env[62952]: DEBUG nova.virt.hardware [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 655.921810] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdea8a33-6b8c-4034-88ff-b57cadb53ff0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.928661] env[62952]: DEBUG nova.compute.manager [req-278d4272-20a3-4661-b021-52d578bfffcb req-5fa967d5-6c81-402d-ab78-11b6aa772d2a service nova] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Received event network-changed-270db801-02b2-419b-9df9-eb2c926b8e24 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 655.928838] env[62952]: DEBUG nova.compute.manager [req-278d4272-20a3-4661-b021-52d578bfffcb req-5fa967d5-6c81-402d-ab78-11b6aa772d2a service nova] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Refreshing instance network info cache due to event network-changed-270db801-02b2-419b-9df9-eb2c926b8e24. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 655.929061] env[62952]: DEBUG oslo_concurrency.lockutils [req-278d4272-20a3-4661-b021-52d578bfffcb req-5fa967d5-6c81-402d-ab78-11b6aa772d2a service nova] Acquiring lock "refresh_cache-25e10d02-8956-4d53-a1a4-b977da38f1c8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.929202] env[62952]: DEBUG oslo_concurrency.lockutils [req-278d4272-20a3-4661-b021-52d578bfffcb req-5fa967d5-6c81-402d-ab78-11b6aa772d2a service nova] Acquired lock "refresh_cache-25e10d02-8956-4d53-a1a4-b977da38f1c8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.929350] env[62952]: DEBUG nova.network.neutron [req-278d4272-20a3-4661-b021-52d578bfffcb req-5fa967d5-6c81-402d-ab78-11b6aa772d2a service nova] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Refreshing network info cache for port 270db801-02b2-419b-9df9-eb2c926b8e24 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 655.935617] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7941816-b50c-4733-a517-2a3139220271 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.042075] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.167135] env[62952]: ERROR nova.compute.manager [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 270db801-02b2-419b-9df9-eb2c926b8e24, please check neutron logs for more information. [ 656.167135] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 656.167135] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.167135] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 656.167135] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.167135] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 656.167135] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.167135] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 656.167135] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.167135] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 656.167135] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.167135] env[62952]: ERROR nova.compute.manager raise self.value [ 656.167135] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.167135] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 656.167135] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.167135] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 656.167630] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.167630] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 656.167630] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 270db801-02b2-419b-9df9-eb2c926b8e24, please check neutron logs for more information. [ 656.167630] env[62952]: ERROR nova.compute.manager [ 656.167630] env[62952]: Traceback (most recent call last): [ 656.167630] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 656.167630] env[62952]: listener.cb(fileno) [ 656.167630] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.167630] env[62952]: result = function(*args, **kwargs) [ 656.167630] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.167630] env[62952]: return func(*args, **kwargs) [ 656.167630] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.167630] env[62952]: raise e [ 656.167630] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.167630] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 656.167630] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.167630] env[62952]: created_port_ids = self._update_ports_for_instance( [ 656.167630] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.167630] env[62952]: with excutils.save_and_reraise_exception(): [ 656.167630] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.167630] env[62952]: self.force_reraise() [ 656.167630] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.167630] env[62952]: raise self.value [ 656.167630] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.167630] env[62952]: updated_port = self._update_port( [ 656.167630] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.167630] env[62952]: _ensure_no_port_binding_failure(port) [ 656.167630] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.167630] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 656.168531] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 270db801-02b2-419b-9df9-eb2c926b8e24, please check neutron logs for more information. [ 656.168531] env[62952]: Removing descriptor: 15 [ 656.168531] env[62952]: ERROR nova.compute.manager [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 270db801-02b2-419b-9df9-eb2c926b8e24, please check neutron logs for more information. [ 656.168531] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Traceback (most recent call last): [ 656.168531] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 656.168531] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] yield resources [ 656.168531] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.168531] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] self.driver.spawn(context, instance, image_meta, [ 656.168531] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 656.168531] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.168531] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.168531] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] vm_ref = self.build_virtual_machine(instance, [ 656.168938] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.168938] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.168938] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.168938] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] for vif in network_info: [ 656.168938] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.168938] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] return self._sync_wrapper(fn, *args, **kwargs) [ 656.168938] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.168938] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] self.wait() [ 656.168938] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.168938] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] self[:] = self._gt.wait() [ 656.168938] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.168938] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] return self._exit_event.wait() [ 656.168938] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 656.169357] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] result = hub.switch() [ 656.169357] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 656.169357] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] return self.greenlet.switch() [ 656.169357] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.169357] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] result = function(*args, **kwargs) [ 656.169357] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.169357] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] return func(*args, **kwargs) [ 656.169357] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.169357] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] raise e [ 656.169357] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.169357] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] nwinfo = self.network_api.allocate_for_instance( [ 656.169357] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.169357] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] created_port_ids = self._update_ports_for_instance( [ 656.169835] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.169835] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] with excutils.save_and_reraise_exception(): [ 656.169835] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.169835] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] self.force_reraise() [ 656.169835] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.169835] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] raise self.value [ 656.169835] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.169835] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] updated_port = self._update_port( [ 656.169835] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.169835] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] _ensure_no_port_binding_failure(port) [ 656.169835] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.169835] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] raise exception.PortBindingFailed(port_id=port['id']) [ 656.170246] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] nova.exception.PortBindingFailed: Binding failed for port 270db801-02b2-419b-9df9-eb2c926b8e24, please check neutron logs for more information. [ 656.170246] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] [ 656.170246] env[62952]: INFO nova.compute.manager [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Terminating instance [ 656.171273] env[62952]: DEBUG oslo_concurrency.lockutils [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Acquiring lock "refresh_cache-25e10d02-8956-4d53-a1a4-b977da38f1c8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.390800] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.536s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.391310] env[62952]: DEBUG nova.compute.manager [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 656.394463] env[62952]: DEBUG oslo_concurrency.lockutils [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.183s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.396051] env[62952]: INFO nova.compute.claims [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 656.458886] env[62952]: DEBUG nova.network.neutron [req-278d4272-20a3-4661-b021-52d578bfffcb req-5fa967d5-6c81-402d-ab78-11b6aa772d2a service nova] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 656.541252] env[62952]: DEBUG nova.network.neutron [req-278d4272-20a3-4661-b021-52d578bfffcb req-5fa967d5-6c81-402d-ab78-11b6aa772d2a service nova] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.907456] env[62952]: DEBUG nova.compute.utils [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 656.908854] env[62952]: DEBUG nova.compute.manager [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 656.909030] env[62952]: DEBUG nova.network.neutron [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 656.968158] env[62952]: DEBUG nova.policy [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae9497b02c4541f1975667d96cbc07fc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b0130ba27a346689de1b666b3da2e59', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 657.045523] env[62952]: DEBUG oslo_concurrency.lockutils [req-278d4272-20a3-4661-b021-52d578bfffcb req-5fa967d5-6c81-402d-ab78-11b6aa772d2a service nova] Releasing lock "refresh_cache-25e10d02-8956-4d53-a1a4-b977da38f1c8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.045941] env[62952]: DEBUG oslo_concurrency.lockutils [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Acquired lock "refresh_cache-25e10d02-8956-4d53-a1a4-b977da38f1c8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.046142] env[62952]: DEBUG nova.network.neutron [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 657.288184] env[62952]: DEBUG nova.network.neutron [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Successfully created port: e2f7d15e-3a04-4206-ab85-de44810e6577 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 657.414169] env[62952]: DEBUG nova.compute.manager [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 657.569886] env[62952]: DEBUG nova.network.neutron [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.646325] env[62952]: DEBUG nova.network.neutron [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.924345] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1663aa05-f8d3-49ce-9fcd-81315f0bd19f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.934545] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e77446-9d54-4fc3-a644-165cbcb52cf1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.973105] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40005521-18ee-4493-839e-2942f23190eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.982718] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c30531-0ca8-432f-9dab-0ed4ff3fc816 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.005268] env[62952]: DEBUG nova.compute.provider_tree [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.149262] env[62952]: DEBUG oslo_concurrency.lockutils [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Releasing lock "refresh_cache-25e10d02-8956-4d53-a1a4-b977da38f1c8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.149666] env[62952]: DEBUG nova.compute.manager [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 658.149882] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 658.150217] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0b68844a-8920-4c3d-a17d-32f4746b0680 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.161959] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6171392a-840b-4afb-93b8-20e171737c73 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.187107] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 25e10d02-8956-4d53-a1a4-b977da38f1c8 could not be found. [ 658.187358] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 658.187541] env[62952]: INFO nova.compute.manager [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 658.187788] env[62952]: DEBUG oslo.service.loopingcall [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 658.188971] env[62952]: DEBUG nova.compute.manager [-] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.189216] env[62952]: DEBUG nova.network.neutron [-] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 658.192226] env[62952]: DEBUG nova.compute.manager [req-1dda469d-222d-4454-bd86-a40fe6b256bc req-04e1758a-e7f1-452c-99c7-fa9f76a8c840 service nova] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Received event network-vif-deleted-270db801-02b2-419b-9df9-eb2c926b8e24 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 658.212549] env[62952]: DEBUG nova.network.neutron [-] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.276699] env[62952]: ERROR nova.compute.manager [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e2f7d15e-3a04-4206-ab85-de44810e6577, please check neutron logs for more information. [ 658.276699] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 658.276699] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.276699] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 658.276699] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 658.276699] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 658.276699] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 658.276699] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 658.276699] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.276699] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 658.276699] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.276699] env[62952]: ERROR nova.compute.manager raise self.value [ 658.276699] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 658.276699] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 658.276699] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.276699] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 658.277544] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.277544] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 658.277544] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e2f7d15e-3a04-4206-ab85-de44810e6577, please check neutron logs for more information. [ 658.277544] env[62952]: ERROR nova.compute.manager [ 658.277544] env[62952]: Traceback (most recent call last): [ 658.277544] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 658.277544] env[62952]: listener.cb(fileno) [ 658.277544] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.277544] env[62952]: result = function(*args, **kwargs) [ 658.277544] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 658.277544] env[62952]: return func(*args, **kwargs) [ 658.277544] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.277544] env[62952]: raise e [ 658.277544] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.277544] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 658.277544] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 658.277544] env[62952]: created_port_ids = self._update_ports_for_instance( [ 658.277544] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 658.277544] env[62952]: with excutils.save_and_reraise_exception(): [ 658.277544] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.277544] env[62952]: self.force_reraise() [ 658.277544] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.277544] env[62952]: raise self.value [ 658.277544] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 658.277544] env[62952]: updated_port = self._update_port( [ 658.277544] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.277544] env[62952]: _ensure_no_port_binding_failure(port) [ 658.277544] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.277544] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 658.279055] env[62952]: nova.exception.PortBindingFailed: Binding failed for port e2f7d15e-3a04-4206-ab85-de44810e6577, please check neutron logs for more information. [ 658.279055] env[62952]: Removing descriptor: 15 [ 658.429962] env[62952]: DEBUG nova.compute.manager [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 658.458024] env[62952]: DEBUG nova.virt.hardware [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 658.458351] env[62952]: DEBUG nova.virt.hardware [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 658.458568] env[62952]: DEBUG nova.virt.hardware [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 658.458812] env[62952]: DEBUG nova.virt.hardware [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 658.458976] env[62952]: DEBUG nova.virt.hardware [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 658.459276] env[62952]: DEBUG nova.virt.hardware [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 658.459504] env[62952]: DEBUG nova.virt.hardware [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 658.459710] env[62952]: DEBUG nova.virt.hardware [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 658.459971] env[62952]: DEBUG nova.virt.hardware [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 658.460167] env[62952]: DEBUG nova.virt.hardware [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 658.460400] env[62952]: DEBUG nova.virt.hardware [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 658.461257] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ef6918-3895-4cb9-9fec-668e3c40632d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.469482] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38284d17-b48c-4133-9c4e-ab502e5e2e3b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.484422] env[62952]: ERROR nova.compute.manager [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e2f7d15e-3a04-4206-ab85-de44810e6577, please check neutron logs for more information. [ 658.484422] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Traceback (most recent call last): [ 658.484422] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 658.484422] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] yield resources [ 658.484422] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 658.484422] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] self.driver.spawn(context, instance, image_meta, [ 658.484422] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 658.484422] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 658.484422] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 658.484422] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] vm_ref = self.build_virtual_machine(instance, [ 658.484422] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 658.484894] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] vif_infos = vmwarevif.get_vif_info(self._session, [ 658.484894] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 658.484894] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] for vif in network_info: [ 658.484894] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 658.484894] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] return self._sync_wrapper(fn, *args, **kwargs) [ 658.484894] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 658.484894] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] self.wait() [ 658.484894] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 658.484894] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] self[:] = self._gt.wait() [ 658.484894] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 658.484894] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] return self._exit_event.wait() [ 658.484894] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 658.484894] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] current.throw(*self._exc) [ 658.485390] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.485390] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] result = function(*args, **kwargs) [ 658.485390] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 658.485390] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] return func(*args, **kwargs) [ 658.485390] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.485390] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] raise e [ 658.485390] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.485390] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] nwinfo = self.network_api.allocate_for_instance( [ 658.485390] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 658.485390] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] created_port_ids = self._update_ports_for_instance( [ 658.485390] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 658.485390] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] with excutils.save_and_reraise_exception(): [ 658.485390] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.485825] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] self.force_reraise() [ 658.485825] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.485825] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] raise self.value [ 658.485825] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 658.485825] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] updated_port = self._update_port( [ 658.485825] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.485825] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] _ensure_no_port_binding_failure(port) [ 658.485825] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.485825] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] raise exception.PortBindingFailed(port_id=port['id']) [ 658.485825] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] nova.exception.PortBindingFailed: Binding failed for port e2f7d15e-3a04-4206-ab85-de44810e6577, please check neutron logs for more information. [ 658.485825] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] [ 658.485825] env[62952]: INFO nova.compute.manager [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Terminating instance [ 658.487014] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Acquiring lock "refresh_cache-78a0acf9-12f5-4699-a4b8-2cbefcbd343b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.487187] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Acquired lock "refresh_cache-78a0acf9-12f5-4699-a4b8-2cbefcbd343b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.487350] env[62952]: DEBUG nova.network.neutron [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 658.508945] env[62952]: DEBUG nova.scheduler.client.report [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 658.712354] env[62952]: DEBUG nova.network.neutron [-] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.014175] env[62952]: DEBUG nova.network.neutron [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 659.017805] env[62952]: DEBUG oslo_concurrency.lockutils [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.623s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.018477] env[62952]: DEBUG nova.compute.manager [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 659.020987] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.951s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.024633] env[62952]: INFO nova.compute.claims [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 659.155246] env[62952]: DEBUG nova.network.neutron [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.215210] env[62952]: INFO nova.compute.manager [-] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Took 1.03 seconds to deallocate network for instance. [ 659.218310] env[62952]: DEBUG nova.compute.claims [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 659.218737] env[62952]: DEBUG oslo_concurrency.lockutils [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.533125] env[62952]: DEBUG nova.compute.utils [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 659.534177] env[62952]: DEBUG nova.compute.manager [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 659.534499] env[62952]: DEBUG nova.network.neutron [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 659.599014] env[62952]: DEBUG nova.policy [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de18659bccf44676b60a42308cac7b72', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d16d9b46fbe4012b533053092922b23', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 659.666163] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Releasing lock "refresh_cache-78a0acf9-12f5-4699-a4b8-2cbefcbd343b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.671354] env[62952]: DEBUG nova.compute.manager [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 659.671354] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 659.671354] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c64fa5c6-7624-415b-85cb-8ac08b927fbb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.684224] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23d433c1-8c9d-490d-881b-9f4db27ffdcd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.712677] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 78a0acf9-12f5-4699-a4b8-2cbefcbd343b could not be found. [ 659.712906] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 659.713099] env[62952]: INFO nova.compute.manager [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 659.713432] env[62952]: DEBUG oslo.service.loopingcall [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 659.713693] env[62952]: DEBUG nova.compute.manager [-] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 659.713785] env[62952]: DEBUG nova.network.neutron [-] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 659.752772] env[62952]: DEBUG nova.network.neutron [-] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 660.040359] env[62952]: DEBUG nova.network.neutron [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Successfully created port: f3a6797c-20d9-4a73-b4ef-f147a2a28c20 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 660.042453] env[62952]: DEBUG nova.compute.manager [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 660.220069] env[62952]: DEBUG nova.compute.manager [req-495a1735-fa00-48f0-88e0-ac7e6c4fd3ca req-58584c90-0dcf-4efd-b7eb-1767c2461b58 service nova] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Received event network-changed-e2f7d15e-3a04-4206-ab85-de44810e6577 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 660.223795] env[62952]: DEBUG nova.compute.manager [req-495a1735-fa00-48f0-88e0-ac7e6c4fd3ca req-58584c90-0dcf-4efd-b7eb-1767c2461b58 service nova] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Refreshing instance network info cache due to event network-changed-e2f7d15e-3a04-4206-ab85-de44810e6577. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 660.224037] env[62952]: DEBUG oslo_concurrency.lockutils [req-495a1735-fa00-48f0-88e0-ac7e6c4fd3ca req-58584c90-0dcf-4efd-b7eb-1767c2461b58 service nova] Acquiring lock "refresh_cache-78a0acf9-12f5-4699-a4b8-2cbefcbd343b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.224187] env[62952]: DEBUG oslo_concurrency.lockutils [req-495a1735-fa00-48f0-88e0-ac7e6c4fd3ca req-58584c90-0dcf-4efd-b7eb-1767c2461b58 service nova] Acquired lock "refresh_cache-78a0acf9-12f5-4699-a4b8-2cbefcbd343b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.224345] env[62952]: DEBUG nova.network.neutron [req-495a1735-fa00-48f0-88e0-ac7e6c4fd3ca req-58584c90-0dcf-4efd-b7eb-1767c2461b58 service nova] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Refreshing network info cache for port e2f7d15e-3a04-4206-ab85-de44810e6577 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 660.257892] env[62952]: DEBUG nova.network.neutron [-] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.435902] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3f6001-41dc-4ad2-a00a-23172ee2a6e3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.444954] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e17d55-8e9c-4338-9fa2-0add2caea335 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.477501] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659212fd-f86f-4300-919f-21899ff8dd3c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.486441] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f89f33a-77d7-40f3-810e-bb67c116e296 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.500946] env[62952]: DEBUG nova.compute.provider_tree [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.747610] env[62952]: DEBUG nova.network.neutron [req-495a1735-fa00-48f0-88e0-ac7e6c4fd3ca req-58584c90-0dcf-4efd-b7eb-1767c2461b58 service nova] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 660.762067] env[62952]: INFO nova.compute.manager [-] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Took 1.05 seconds to deallocate network for instance. [ 660.765274] env[62952]: DEBUG nova.compute.claims [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 660.765566] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.889406] env[62952]: DEBUG nova.network.neutron [req-495a1735-fa00-48f0-88e0-ac7e6c4fd3ca req-58584c90-0dcf-4efd-b7eb-1767c2461b58 service nova] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.004486] env[62952]: DEBUG nova.scheduler.client.report [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.043649] env[62952]: ERROR nova.compute.manager [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f3a6797c-20d9-4a73-b4ef-f147a2a28c20, please check neutron logs for more information. [ 661.043649] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 661.043649] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.043649] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 661.043649] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.043649] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 661.043649] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.043649] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 661.043649] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.043649] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 661.043649] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.043649] env[62952]: ERROR nova.compute.manager raise self.value [ 661.043649] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.043649] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 661.043649] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.043649] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 661.044495] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.044495] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 661.044495] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f3a6797c-20d9-4a73-b4ef-f147a2a28c20, please check neutron logs for more information. [ 661.044495] env[62952]: ERROR nova.compute.manager [ 661.044495] env[62952]: Traceback (most recent call last): [ 661.044495] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 661.044495] env[62952]: listener.cb(fileno) [ 661.044495] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.044495] env[62952]: result = function(*args, **kwargs) [ 661.044495] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.044495] env[62952]: return func(*args, **kwargs) [ 661.044495] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.044495] env[62952]: raise e [ 661.044495] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.044495] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 661.044495] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.044495] env[62952]: created_port_ids = self._update_ports_for_instance( [ 661.044495] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.044495] env[62952]: with excutils.save_and_reraise_exception(): [ 661.044495] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.044495] env[62952]: self.force_reraise() [ 661.044495] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.044495] env[62952]: raise self.value [ 661.044495] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.044495] env[62952]: updated_port = self._update_port( [ 661.044495] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.044495] env[62952]: _ensure_no_port_binding_failure(port) [ 661.044495] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.044495] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 661.045520] env[62952]: nova.exception.PortBindingFailed: Binding failed for port f3a6797c-20d9-4a73-b4ef-f147a2a28c20, please check neutron logs for more information. [ 661.045520] env[62952]: Removing descriptor: 15 [ 661.062510] env[62952]: DEBUG nova.compute.manager [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 661.088985] env[62952]: DEBUG nova.virt.hardware [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 661.090748] env[62952]: DEBUG nova.virt.hardware [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 661.090748] env[62952]: DEBUG nova.virt.hardware [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 661.090748] env[62952]: DEBUG nova.virt.hardware [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 661.090748] env[62952]: DEBUG nova.virt.hardware [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 661.090748] env[62952]: DEBUG nova.virt.hardware [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 661.090957] env[62952]: DEBUG nova.virt.hardware [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 661.090957] env[62952]: DEBUG nova.virt.hardware [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 661.090957] env[62952]: DEBUG nova.virt.hardware [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 661.090957] env[62952]: DEBUG nova.virt.hardware [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 661.090957] env[62952]: DEBUG nova.virt.hardware [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 661.091971] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b494fdb-87fa-4637-8a8d-78148b04e279 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.101802] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03bfc5e-6074-459f-8c8c-d47213e62c04 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.118067] env[62952]: ERROR nova.compute.manager [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f3a6797c-20d9-4a73-b4ef-f147a2a28c20, please check neutron logs for more information. [ 661.118067] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Traceback (most recent call last): [ 661.118067] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 661.118067] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] yield resources [ 661.118067] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 661.118067] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] self.driver.spawn(context, instance, image_meta, [ 661.118067] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 661.118067] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.118067] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.118067] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] vm_ref = self.build_virtual_machine(instance, [ 661.118067] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.119098] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.119098] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.119098] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] for vif in network_info: [ 661.119098] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 661.119098] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] return self._sync_wrapper(fn, *args, **kwargs) [ 661.119098] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 661.119098] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] self.wait() [ 661.119098] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 661.119098] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] self[:] = self._gt.wait() [ 661.119098] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.119098] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] return self._exit_event.wait() [ 661.119098] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 661.119098] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] current.throw(*self._exc) [ 661.119687] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.119687] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] result = function(*args, **kwargs) [ 661.119687] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.119687] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] return func(*args, **kwargs) [ 661.119687] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.119687] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] raise e [ 661.119687] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.119687] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] nwinfo = self.network_api.allocate_for_instance( [ 661.119687] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.119687] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] created_port_ids = self._update_ports_for_instance( [ 661.119687] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.119687] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] with excutils.save_and_reraise_exception(): [ 661.119687] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.121047] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] self.force_reraise() [ 661.121047] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.121047] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] raise self.value [ 661.121047] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.121047] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] updated_port = self._update_port( [ 661.121047] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.121047] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] _ensure_no_port_binding_failure(port) [ 661.121047] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.121047] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] raise exception.PortBindingFailed(port_id=port['id']) [ 661.121047] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] nova.exception.PortBindingFailed: Binding failed for port f3a6797c-20d9-4a73-b4ef-f147a2a28c20, please check neutron logs for more information. [ 661.121047] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] [ 661.121047] env[62952]: INFO nova.compute.manager [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Terminating instance [ 661.121439] env[62952]: DEBUG oslo_concurrency.lockutils [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Acquiring lock "refresh_cache-4eb9ea32-b2eb-440c-871b-6942590d9c01" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.121439] env[62952]: DEBUG oslo_concurrency.lockutils [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Acquired lock "refresh_cache-4eb9ea32-b2eb-440c-871b-6942590d9c01" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.121516] env[62952]: DEBUG nova.network.neutron [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 661.392499] env[62952]: DEBUG oslo_concurrency.lockutils [req-495a1735-fa00-48f0-88e0-ac7e6c4fd3ca req-58584c90-0dcf-4efd-b7eb-1767c2461b58 service nova] Releasing lock "refresh_cache-78a0acf9-12f5-4699-a4b8-2cbefcbd343b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.392726] env[62952]: DEBUG nova.compute.manager [req-495a1735-fa00-48f0-88e0-ac7e6c4fd3ca req-58584c90-0dcf-4efd-b7eb-1767c2461b58 service nova] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Received event network-vif-deleted-e2f7d15e-3a04-4206-ab85-de44810e6577 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 661.511130] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.490s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.511130] env[62952]: DEBUG nova.compute.manager [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 661.514106] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.793s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.518381] env[62952]: INFO nova.compute.claims [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 661.644700] env[62952]: DEBUG nova.network.neutron [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.770908] env[62952]: DEBUG nova.network.neutron [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.022318] env[62952]: DEBUG nova.compute.utils [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 662.025753] env[62952]: DEBUG nova.compute.manager [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 662.025933] env[62952]: DEBUG nova.network.neutron [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 662.078888] env[62952]: DEBUG nova.policy [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '65821aa0b1224fcf8252fe74f1a49301', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '650857e639404b50afe86add5a9b487c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 662.170991] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Acquiring lock "3822c92d-560b-4d61-92e9-7c694bc7a33a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.171267] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Lock "3822c92d-560b-4d61-92e9-7c694bc7a33a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.251067] env[62952]: DEBUG nova.compute.manager [req-1741d68c-2536-4417-b254-9c8d221c0fd0 req-b9856df8-630b-43a2-b95d-2a13fda3d61a service nova] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Received event network-changed-f3a6797c-20d9-4a73-b4ef-f147a2a28c20 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 662.251265] env[62952]: DEBUG nova.compute.manager [req-1741d68c-2536-4417-b254-9c8d221c0fd0 req-b9856df8-630b-43a2-b95d-2a13fda3d61a service nova] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Refreshing instance network info cache due to event network-changed-f3a6797c-20d9-4a73-b4ef-f147a2a28c20. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 662.251452] env[62952]: DEBUG oslo_concurrency.lockutils [req-1741d68c-2536-4417-b254-9c8d221c0fd0 req-b9856df8-630b-43a2-b95d-2a13fda3d61a service nova] Acquiring lock "refresh_cache-4eb9ea32-b2eb-440c-871b-6942590d9c01" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.277911] env[62952]: DEBUG oslo_concurrency.lockutils [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Releasing lock "refresh_cache-4eb9ea32-b2eb-440c-871b-6942590d9c01" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.278405] env[62952]: DEBUG nova.compute.manager [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 662.279552] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 662.279552] env[62952]: DEBUG oslo_concurrency.lockutils [req-1741d68c-2536-4417-b254-9c8d221c0fd0 req-b9856df8-630b-43a2-b95d-2a13fda3d61a service nova] Acquired lock "refresh_cache-4eb9ea32-b2eb-440c-871b-6942590d9c01" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.279552] env[62952]: DEBUG nova.network.neutron [req-1741d68c-2536-4417-b254-9c8d221c0fd0 req-b9856df8-630b-43a2-b95d-2a13fda3d61a service nova] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Refreshing network info cache for port f3a6797c-20d9-4a73-b4ef-f147a2a28c20 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 662.281228] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b5fe4efe-e8f3-4b10-8c50-13957cb67d1b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.291300] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b375a55b-10cd-404f-a3a3-15f5bda4d846 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.316866] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4eb9ea32-b2eb-440c-871b-6942590d9c01 could not be found. [ 662.317391] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 662.317696] env[62952]: INFO nova.compute.manager [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Took 0.04 seconds to destroy the instance on the hypervisor. [ 662.318080] env[62952]: DEBUG oslo.service.loopingcall [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 662.318529] env[62952]: DEBUG nova.compute.manager [-] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 662.318744] env[62952]: DEBUG nova.network.neutron [-] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 662.342878] env[62952]: DEBUG nova.network.neutron [-] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.389014] env[62952]: DEBUG nova.network.neutron [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Successfully created port: e46c02d0-3a45-4099-8006-a529a1c1f2df {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 662.527642] env[62952]: DEBUG nova.compute.manager [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 662.803992] env[62952]: DEBUG nova.network.neutron [req-1741d68c-2536-4417-b254-9c8d221c0fd0 req-b9856df8-630b-43a2-b95d-2a13fda3d61a service nova] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.845879] env[62952]: DEBUG nova.network.neutron [-] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.931063] env[62952]: DEBUG nova.network.neutron [req-1741d68c-2536-4417-b254-9c8d221c0fd0 req-b9856df8-630b-43a2-b95d-2a13fda3d61a service nova] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.999830] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0018d262-7e2c-47fb-8969-4a2636e0ead0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.010308] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b3e72f-e77e-4717-bcad-b2e522cb070c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.045565] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b3ecf5-9c67-425f-9b35-2aac6c574f1a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.054850] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32820581-0202-4134-95c5-f032b4c4403c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.070242] env[62952]: DEBUG nova.compute.provider_tree [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.342261] env[62952]: ERROR nova.compute.manager [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e46c02d0-3a45-4099-8006-a529a1c1f2df, please check neutron logs for more information. [ 663.342261] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 663.342261] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.342261] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 663.342261] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.342261] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 663.342261] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.342261] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 663.342261] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.342261] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 663.342261] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.342261] env[62952]: ERROR nova.compute.manager raise self.value [ 663.342261] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.342261] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 663.342261] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.342261] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 663.343061] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.343061] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 663.343061] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e46c02d0-3a45-4099-8006-a529a1c1f2df, please check neutron logs for more information. [ 663.343061] env[62952]: ERROR nova.compute.manager [ 663.343061] env[62952]: Traceback (most recent call last): [ 663.343061] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 663.343061] env[62952]: listener.cb(fileno) [ 663.343061] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.343061] env[62952]: result = function(*args, **kwargs) [ 663.343061] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.343061] env[62952]: return func(*args, **kwargs) [ 663.343061] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.343061] env[62952]: raise e [ 663.343061] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.343061] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 663.343061] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.343061] env[62952]: created_port_ids = self._update_ports_for_instance( [ 663.343061] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.343061] env[62952]: with excutils.save_and_reraise_exception(): [ 663.343061] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.343061] env[62952]: self.force_reraise() [ 663.343061] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.343061] env[62952]: raise self.value [ 663.343061] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.343061] env[62952]: updated_port = self._update_port( [ 663.343061] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.343061] env[62952]: _ensure_no_port_binding_failure(port) [ 663.343061] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.343061] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 663.344018] env[62952]: nova.exception.PortBindingFailed: Binding failed for port e46c02d0-3a45-4099-8006-a529a1c1f2df, please check neutron logs for more information. [ 663.344018] env[62952]: Removing descriptor: 15 [ 663.349360] env[62952]: INFO nova.compute.manager [-] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Took 1.03 seconds to deallocate network for instance. [ 663.352023] env[62952]: DEBUG nova.compute.claims [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 663.352282] env[62952]: DEBUG oslo_concurrency.lockutils [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.433140] env[62952]: DEBUG oslo_concurrency.lockutils [req-1741d68c-2536-4417-b254-9c8d221c0fd0 req-b9856df8-630b-43a2-b95d-2a13fda3d61a service nova] Releasing lock "refresh_cache-4eb9ea32-b2eb-440c-871b-6942590d9c01" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.433594] env[62952]: DEBUG nova.compute.manager [req-1741d68c-2536-4417-b254-9c8d221c0fd0 req-b9856df8-630b-43a2-b95d-2a13fda3d61a service nova] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Received event network-vif-deleted-f3a6797c-20d9-4a73-b4ef-f147a2a28c20 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 663.550191] env[62952]: DEBUG nova.compute.manager [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 663.575952] env[62952]: DEBUG nova.scheduler.client.report [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 663.581071] env[62952]: DEBUG nova.virt.hardware [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 663.581320] env[62952]: DEBUG nova.virt.hardware [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 663.581475] env[62952]: DEBUG nova.virt.hardware [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 663.581651] env[62952]: DEBUG nova.virt.hardware [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 663.581794] env[62952]: DEBUG nova.virt.hardware [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 663.581939] env[62952]: DEBUG nova.virt.hardware [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 663.582154] env[62952]: DEBUG nova.virt.hardware [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 663.582311] env[62952]: DEBUG nova.virt.hardware [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 663.582475] env[62952]: DEBUG nova.virt.hardware [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 663.582634] env[62952]: DEBUG nova.virt.hardware [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 663.582802] env[62952]: DEBUG nova.virt.hardware [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 663.584076] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d17101-0cfd-4482-9dc8-ec8d0be540de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.593376] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1763b8c-8f0f-47bd-82f3-95323a27b595 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.608870] env[62952]: ERROR nova.compute.manager [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e46c02d0-3a45-4099-8006-a529a1c1f2df, please check neutron logs for more information. [ 663.608870] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Traceback (most recent call last): [ 663.608870] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 663.608870] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] yield resources [ 663.608870] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 663.608870] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] self.driver.spawn(context, instance, image_meta, [ 663.608870] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 663.608870] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.608870] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.608870] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] vm_ref = self.build_virtual_machine(instance, [ 663.608870] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.609283] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.609283] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.609283] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] for vif in network_info: [ 663.609283] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.609283] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] return self._sync_wrapper(fn, *args, **kwargs) [ 663.609283] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.609283] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] self.wait() [ 663.609283] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.609283] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] self[:] = self._gt.wait() [ 663.609283] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.609283] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] return self._exit_event.wait() [ 663.609283] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 663.609283] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] current.throw(*self._exc) [ 663.610302] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.610302] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] result = function(*args, **kwargs) [ 663.610302] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.610302] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] return func(*args, **kwargs) [ 663.610302] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.610302] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] raise e [ 663.610302] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.610302] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] nwinfo = self.network_api.allocate_for_instance( [ 663.610302] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.610302] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] created_port_ids = self._update_ports_for_instance( [ 663.610302] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.610302] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] with excutils.save_and_reraise_exception(): [ 663.610302] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.611039] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] self.force_reraise() [ 663.611039] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.611039] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] raise self.value [ 663.611039] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.611039] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] updated_port = self._update_port( [ 663.611039] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.611039] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] _ensure_no_port_binding_failure(port) [ 663.611039] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.611039] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] raise exception.PortBindingFailed(port_id=port['id']) [ 663.611039] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] nova.exception.PortBindingFailed: Binding failed for port e46c02d0-3a45-4099-8006-a529a1c1f2df, please check neutron logs for more information. [ 663.611039] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] [ 663.611039] env[62952]: INFO nova.compute.manager [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Terminating instance [ 663.611639] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Acquiring lock "refresh_cache-f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.611639] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Acquired lock "refresh_cache-f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.611639] env[62952]: DEBUG nova.network.neutron [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 664.089142] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.575s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.089686] env[62952]: DEBUG nova.compute.manager [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 664.092484] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.401s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.093915] env[62952]: INFO nova.compute.claims [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 664.134512] env[62952]: DEBUG nova.network.neutron [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.208348] env[62952]: DEBUG nova.network.neutron [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.414758] env[62952]: DEBUG nova.compute.manager [req-0c894bea-2447-4687-9fca-48e1f7a44e0e req-b95513c3-d3b7-4dc0-895d-51b3f5fb6f7a service nova] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Received event network-changed-e46c02d0-3a45-4099-8006-a529a1c1f2df {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 664.415237] env[62952]: DEBUG nova.compute.manager [req-0c894bea-2447-4687-9fca-48e1f7a44e0e req-b95513c3-d3b7-4dc0-895d-51b3f5fb6f7a service nova] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Refreshing instance network info cache due to event network-changed-e46c02d0-3a45-4099-8006-a529a1c1f2df. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 664.415237] env[62952]: DEBUG oslo_concurrency.lockutils [req-0c894bea-2447-4687-9fca-48e1f7a44e0e req-b95513c3-d3b7-4dc0-895d-51b3f5fb6f7a service nova] Acquiring lock "refresh_cache-f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.603903] env[62952]: DEBUG nova.compute.utils [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 664.609573] env[62952]: DEBUG nova.compute.manager [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 664.609573] env[62952]: DEBUG nova.network.neutron [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 664.665632] env[62952]: DEBUG nova.policy [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bbb9fc36c20a46ea866db9291e1df9f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ef331636cec4936a90fb5bacf9bfa39', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 664.712792] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Releasing lock "refresh_cache-f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.713977] env[62952]: DEBUG nova.compute.manager [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 664.713977] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 664.713977] env[62952]: DEBUG oslo_concurrency.lockutils [req-0c894bea-2447-4687-9fca-48e1f7a44e0e req-b95513c3-d3b7-4dc0-895d-51b3f5fb6f7a service nova] Acquired lock "refresh_cache-f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.713977] env[62952]: DEBUG nova.network.neutron [req-0c894bea-2447-4687-9fca-48e1f7a44e0e req-b95513c3-d3b7-4dc0-895d-51b3f5fb6f7a service nova] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Refreshing network info cache for port e46c02d0-3a45-4099-8006-a529a1c1f2df {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 664.716012] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ffe6eb2-1fe6-4275-8888-3d15e35ac6c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.729911] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03812094-a506-41a5-9783-c9ab42ba2c90 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.756100] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83 could not be found. [ 664.756335] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 664.756512] env[62952]: INFO nova.compute.manager [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Took 0.04 seconds to destroy the instance on the hypervisor. [ 664.756946] env[62952]: DEBUG oslo.service.loopingcall [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 664.757049] env[62952]: DEBUG nova.compute.manager [-] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 664.757156] env[62952]: DEBUG nova.network.neutron [-] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 664.780025] env[62952]: DEBUG nova.network.neutron [-] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.003176] env[62952]: DEBUG nova.network.neutron [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Successfully created port: 2f35b2b4-e110-4bb2-a3c3-0579a983f5a7 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 665.108194] env[62952]: DEBUG nova.compute.manager [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 665.242942] env[62952]: DEBUG nova.network.neutron [req-0c894bea-2447-4687-9fca-48e1f7a44e0e req-b95513c3-d3b7-4dc0-895d-51b3f5fb6f7a service nova] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.283586] env[62952]: DEBUG nova.network.neutron [-] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.358858] env[62952]: DEBUG nova.network.neutron [req-0c894bea-2447-4687-9fca-48e1f7a44e0e req-b95513c3-d3b7-4dc0-895d-51b3f5fb6f7a service nova] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.564515] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f67c70f5-fbd9-4544-993b-6ad7fd3259ac {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.574117] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf7e7091-ffa1-436f-af33-51c00acd7c48 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.607217] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63091723-61e5-4725-a35c-62829f757338 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.619976] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa1e776e-4b0a-41f0-956b-dcdaaaa71d7e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.636220] env[62952]: DEBUG nova.compute.provider_tree [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.786693] env[62952]: INFO nova.compute.manager [-] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Took 1.03 seconds to deallocate network for instance. [ 665.790855] env[62952]: DEBUG nova.compute.claims [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 665.790944] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.862426] env[62952]: DEBUG oslo_concurrency.lockutils [req-0c894bea-2447-4687-9fca-48e1f7a44e0e req-b95513c3-d3b7-4dc0-895d-51b3f5fb6f7a service nova] Releasing lock "refresh_cache-f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.862630] env[62952]: DEBUG nova.compute.manager [req-0c894bea-2447-4687-9fca-48e1f7a44e0e req-b95513c3-d3b7-4dc0-895d-51b3f5fb6f7a service nova] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Received event network-vif-deleted-e46c02d0-3a45-4099-8006-a529a1c1f2df {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 665.926926] env[62952]: ERROR nova.compute.manager [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2f35b2b4-e110-4bb2-a3c3-0579a983f5a7, please check neutron logs for more information. [ 665.926926] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 665.926926] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.926926] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 665.926926] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.926926] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 665.926926] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.926926] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 665.926926] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.926926] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 665.926926] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.926926] env[62952]: ERROR nova.compute.manager raise self.value [ 665.926926] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.926926] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 665.926926] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.926926] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 665.927503] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.927503] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 665.927503] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2f35b2b4-e110-4bb2-a3c3-0579a983f5a7, please check neutron logs for more information. [ 665.927503] env[62952]: ERROR nova.compute.manager [ 665.927503] env[62952]: Traceback (most recent call last): [ 665.927503] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 665.927503] env[62952]: listener.cb(fileno) [ 665.927503] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.927503] env[62952]: result = function(*args, **kwargs) [ 665.927503] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.927503] env[62952]: return func(*args, **kwargs) [ 665.927503] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.927503] env[62952]: raise e [ 665.927503] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.927503] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 665.927503] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.927503] env[62952]: created_port_ids = self._update_ports_for_instance( [ 665.927503] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.927503] env[62952]: with excutils.save_and_reraise_exception(): [ 665.927503] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.927503] env[62952]: self.force_reraise() [ 665.927503] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.927503] env[62952]: raise self.value [ 665.927503] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.927503] env[62952]: updated_port = self._update_port( [ 665.927503] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.927503] env[62952]: _ensure_no_port_binding_failure(port) [ 665.927503] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.927503] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 665.928493] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 2f35b2b4-e110-4bb2-a3c3-0579a983f5a7, please check neutron logs for more information. [ 665.928493] env[62952]: Removing descriptor: 15 [ 666.120722] env[62952]: DEBUG nova.compute.manager [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 666.138716] env[62952]: DEBUG nova.scheduler.client.report [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.149575] env[62952]: DEBUG nova.virt.hardware [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 666.149804] env[62952]: DEBUG nova.virt.hardware [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 666.149996] env[62952]: DEBUG nova.virt.hardware [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 666.150203] env[62952]: DEBUG nova.virt.hardware [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 666.150347] env[62952]: DEBUG nova.virt.hardware [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 666.150488] env[62952]: DEBUG nova.virt.hardware [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 666.150688] env[62952]: DEBUG nova.virt.hardware [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 666.150844] env[62952]: DEBUG nova.virt.hardware [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 666.151012] env[62952]: DEBUG nova.virt.hardware [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 666.151177] env[62952]: DEBUG nova.virt.hardware [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 666.151343] env[62952]: DEBUG nova.virt.hardware [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 666.152204] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fad542-1bd5-4fc0-be01-8c9f8dd90f21 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.161754] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1428e608-3fc4-4f68-942f-42dea6f2dea5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.176439] env[62952]: ERROR nova.compute.manager [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2f35b2b4-e110-4bb2-a3c3-0579a983f5a7, please check neutron logs for more information. [ 666.176439] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Traceback (most recent call last): [ 666.176439] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 666.176439] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] yield resources [ 666.176439] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 666.176439] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] self.driver.spawn(context, instance, image_meta, [ 666.176439] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 666.176439] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 666.176439] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 666.176439] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] vm_ref = self.build_virtual_machine(instance, [ 666.176439] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 666.177078] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] vif_infos = vmwarevif.get_vif_info(self._session, [ 666.177078] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 666.177078] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] for vif in network_info: [ 666.177078] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 666.177078] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] return self._sync_wrapper(fn, *args, **kwargs) [ 666.177078] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 666.177078] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] self.wait() [ 666.177078] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 666.177078] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] self[:] = self._gt.wait() [ 666.177078] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 666.177078] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] return self._exit_event.wait() [ 666.177078] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 666.177078] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] current.throw(*self._exc) [ 666.177466] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.177466] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] result = function(*args, **kwargs) [ 666.177466] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 666.177466] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] return func(*args, **kwargs) [ 666.177466] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.177466] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] raise e [ 666.177466] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.177466] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] nwinfo = self.network_api.allocate_for_instance( [ 666.177466] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 666.177466] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] created_port_ids = self._update_ports_for_instance( [ 666.177466] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 666.177466] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] with excutils.save_and_reraise_exception(): [ 666.177466] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.177907] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] self.force_reraise() [ 666.177907] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.177907] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] raise self.value [ 666.177907] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 666.177907] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] updated_port = self._update_port( [ 666.177907] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.177907] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] _ensure_no_port_binding_failure(port) [ 666.177907] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.177907] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] raise exception.PortBindingFailed(port_id=port['id']) [ 666.177907] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] nova.exception.PortBindingFailed: Binding failed for port 2f35b2b4-e110-4bb2-a3c3-0579a983f5a7, please check neutron logs for more information. [ 666.177907] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] [ 666.177907] env[62952]: INFO nova.compute.manager [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Terminating instance [ 666.179124] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Acquiring lock "refresh_cache-6ac19031-ac4e-4061-bbfd-75bd5c90803a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.179379] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Acquired lock "refresh_cache-6ac19031-ac4e-4061-bbfd-75bd5c90803a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.179452] env[62952]: DEBUG nova.network.neutron [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 666.450795] env[62952]: DEBUG nova.compute.manager [req-e26f521d-b076-4144-859b-7da4efcf06e6 req-1c6fc5f2-de58-453e-a52a-35b8aa9aff3d service nova] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Received event network-changed-2f35b2b4-e110-4bb2-a3c3-0579a983f5a7 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 666.450993] env[62952]: DEBUG nova.compute.manager [req-e26f521d-b076-4144-859b-7da4efcf06e6 req-1c6fc5f2-de58-453e-a52a-35b8aa9aff3d service nova] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Refreshing instance network info cache due to event network-changed-2f35b2b4-e110-4bb2-a3c3-0579a983f5a7. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 666.451195] env[62952]: DEBUG oslo_concurrency.lockutils [req-e26f521d-b076-4144-859b-7da4efcf06e6 req-1c6fc5f2-de58-453e-a52a-35b8aa9aff3d service nova] Acquiring lock "refresh_cache-6ac19031-ac4e-4061-bbfd-75bd5c90803a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.645850] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.553s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.646400] env[62952]: DEBUG nova.compute.manager [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 666.648996] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.860s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.650423] env[62952]: INFO nova.compute.claims [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 666.700982] env[62952]: DEBUG nova.network.neutron [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.804422] env[62952]: DEBUG nova.network.neutron [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.158133] env[62952]: DEBUG nova.compute.utils [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 667.160866] env[62952]: DEBUG nova.compute.manager [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Not allocating networking since 'none' was specified. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 667.306350] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Releasing lock "refresh_cache-6ac19031-ac4e-4061-bbfd-75bd5c90803a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.306552] env[62952]: DEBUG nova.compute.manager [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 667.306724] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 667.307114] env[62952]: DEBUG oslo_concurrency.lockutils [req-e26f521d-b076-4144-859b-7da4efcf06e6 req-1c6fc5f2-de58-453e-a52a-35b8aa9aff3d service nova] Acquired lock "refresh_cache-6ac19031-ac4e-4061-bbfd-75bd5c90803a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.307303] env[62952]: DEBUG nova.network.neutron [req-e26f521d-b076-4144-859b-7da4efcf06e6 req-1c6fc5f2-de58-453e-a52a-35b8aa9aff3d service nova] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Refreshing network info cache for port 2f35b2b4-e110-4bb2-a3c3-0579a983f5a7 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 667.308370] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0b33bcb0-dd55-4cac-8042-0f960a68db7c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.318501] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82aed5ec-5d5a-491f-be4e-7429aea7f50b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.344192] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6ac19031-ac4e-4061-bbfd-75bd5c90803a could not be found. [ 667.348271] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 667.348271] env[62952]: INFO nova.compute.manager [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 667.348271] env[62952]: DEBUG oslo.service.loopingcall [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 667.348271] env[62952]: DEBUG nova.compute.manager [-] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.348271] env[62952]: DEBUG nova.network.neutron [-] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 667.366766] env[62952]: DEBUG nova.network.neutron [-] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.662437] env[62952]: DEBUG nova.compute.manager [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 667.826503] env[62952]: DEBUG nova.network.neutron [req-e26f521d-b076-4144-859b-7da4efcf06e6 req-1c6fc5f2-de58-453e-a52a-35b8aa9aff3d service nova] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.868537] env[62952]: DEBUG nova.network.neutron [-] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.915103] env[62952]: DEBUG nova.network.neutron [req-e26f521d-b076-4144-859b-7da4efcf06e6 req-1c6fc5f2-de58-453e-a52a-35b8aa9aff3d service nova] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.017343] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05869f43-7991-4816-8b7a-64135012c6a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.025062] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ffbd3f3-4be3-4f0b-adfc-5268261ea9eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.054846] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0848131-9008-4c84-a855-fa1011cbdbfc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.061660] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc926df7-be52-4910-9e28-e504a58c7555 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.074621] env[62952]: DEBUG nova.compute.provider_tree [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.370949] env[62952]: INFO nova.compute.manager [-] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Took 1.03 seconds to deallocate network for instance. [ 668.373731] env[62952]: DEBUG nova.compute.claims [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 668.373852] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.417495] env[62952]: DEBUG oslo_concurrency.lockutils [req-e26f521d-b076-4144-859b-7da4efcf06e6 req-1c6fc5f2-de58-453e-a52a-35b8aa9aff3d service nova] Releasing lock "refresh_cache-6ac19031-ac4e-4061-bbfd-75bd5c90803a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.417748] env[62952]: DEBUG nova.compute.manager [req-e26f521d-b076-4144-859b-7da4efcf06e6 req-1c6fc5f2-de58-453e-a52a-35b8aa9aff3d service nova] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Received event network-vif-deleted-2f35b2b4-e110-4bb2-a3c3-0579a983f5a7 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 668.578047] env[62952]: DEBUG nova.scheduler.client.report [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.675524] env[62952]: DEBUG nova.compute.manager [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 668.700801] env[62952]: DEBUG nova.virt.hardware [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 668.700993] env[62952]: DEBUG nova.virt.hardware [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 668.701166] env[62952]: DEBUG nova.virt.hardware [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 668.701343] env[62952]: DEBUG nova.virt.hardware [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 668.701485] env[62952]: DEBUG nova.virt.hardware [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 668.701705] env[62952]: DEBUG nova.virt.hardware [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 668.701920] env[62952]: DEBUG nova.virt.hardware [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 668.702097] env[62952]: DEBUG nova.virt.hardware [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 668.702264] env[62952]: DEBUG nova.virt.hardware [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 668.702423] env[62952]: DEBUG nova.virt.hardware [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 668.702591] env[62952]: DEBUG nova.virt.hardware [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 668.703473] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef0135f-e354-45f5-917b-d667bad2d89f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.711415] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a0d5a8-240c-46d8-b813-211555251361 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.725142] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 668.734237] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 668.734526] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fc97c71f-8736-4dc0-8c5a-68e90366c1a7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.747655] env[62952]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 668.747872] env[62952]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62952) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 668.748190] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Folder already exists: OpenStack. Parent ref: group-v4. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 668.748375] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Creating folder: Project (720328bafb1a4e52bfeb4306a5bb96ed). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 668.748596] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b40a9057-63ad-4086-a5dd-484edd7a0892 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.756788] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Created folder: Project (720328bafb1a4e52bfeb4306a5bb96ed) in parent group-v271811. [ 668.756975] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Creating folder: Instances. Parent ref: group-v271816. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 668.757215] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-41dcc547-68cb-4791-8b1b-7f1f7e5d75c9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.764599] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Created folder: Instances in parent group-v271816. [ 668.764808] env[62952]: DEBUG oslo.service.loopingcall [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 668.764992] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 668.765185] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-10459ad4-e441-4d3a-8fc9-853d1cfff984 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.779907] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 668.779907] env[62952]: value = "task-1263052" [ 668.779907] env[62952]: _type = "Task" [ 668.779907] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.787159] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263052, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.083480] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.084017] env[62952]: DEBUG nova.compute.manager [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 669.086807] env[62952]: DEBUG oslo_concurrency.lockutils [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.634s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.289573] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263052, 'name': CreateVM_Task, 'duration_secs': 0.29437} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.292026] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 669.292026] env[62952]: DEBUG oslo_vmware.service [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-633d61c4-1d66-49f7-afde-bbcd93ab2b50 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.296299] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.296459] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.297118] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 669.297352] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdfd7c38-2ac8-4495-b268-aaabeab21d06 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.301294] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Waiting for the task: (returnval){ [ 669.301294] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]528b18ad-0238-344a-0f8e-4364cfe3d850" [ 669.301294] env[62952]: _type = "Task" [ 669.301294] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.308582] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]528b18ad-0238-344a-0f8e-4364cfe3d850, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.595800] env[62952]: DEBUG nova.compute.utils [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 669.598164] env[62952]: DEBUG nova.compute.manager [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 669.598346] env[62952]: DEBUG nova.network.neutron [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 669.639233] env[62952]: DEBUG nova.policy [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ef72aa6e64d4378883e53894dfae086', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e09d6d5c5fe8472b88bf4e7c3d0d0467', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 669.813917] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.814204] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 669.814388] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.814529] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.815044] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 669.815465] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9e620b56-09cc-46d5-ab27-97bffecd6a25 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.832384] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 669.832627] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 669.833448] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ac383a-aee5-4d02-8f53-5be69cbde4e0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.841860] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-286686b2-b268-4b68-bf79-2ec6957b29b7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.846335] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Waiting for the task: (returnval){ [ 669.846335] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5234d90d-e6bb-3b12-958b-9c2a6ccff22f" [ 669.846335] env[62952]: _type = "Task" [ 669.846335] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.857560] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5234d90d-e6bb-3b12-958b-9c2a6ccff22f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.917148] env[62952]: DEBUG nova.network.neutron [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Successfully created port: 1d9877c9-1496-4654-a090-78b82fabb64e {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 669.990183] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c47383b-5107-485d-a72d-08a79b8a073d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.998119] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e641ac4-dd28-44c8-888e-a8e95bd1fa77 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.028457] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e19e4dc-b53b-4eba-943a-934bd3644e94 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.035781] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52d974c2-7d30-4524-aa13-e23535685eca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.049218] env[62952]: DEBUG nova.compute.provider_tree [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.098673] env[62952]: DEBUG nova.compute.manager [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 670.366896] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Preparing fetch location {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 670.367242] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Creating directory with path [datastore1] vmware_temp/0448dc92-cb44-4a57-b5be-e39566a84600/119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 670.367506] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c1741756-8f6c-4d38-b746-70e72bcdd45d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.388800] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Created directory with path [datastore1] vmware_temp/0448dc92-cb44-4a57-b5be-e39566a84600/119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 670.389063] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Fetch image to [datastore1] vmware_temp/0448dc92-cb44-4a57-b5be-e39566a84600/119dd2b4-b6de-42d5-adba-db818e4c2cd7/tmp-sparse.vmdk {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 670.389239] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Downloading image file data 119dd2b4-b6de-42d5-adba-db818e4c2cd7 to [datastore1] vmware_temp/0448dc92-cb44-4a57-b5be-e39566a84600/119dd2b4-b6de-42d5-adba-db818e4c2cd7/tmp-sparse.vmdk on the data store datastore1 {{(pid=62952) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 670.390365] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f5eec52-d840-4109-8a4f-2292f33fd090 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.399227] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd32c71f-367c-4514-9051-a33be50ad376 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.408872] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720c59a8-821c-4eb0-bfc8-3d577742db36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.440631] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4312f996-e668-4b9e-8d9f-c53815750ca4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.446762] env[62952]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-70648359-4129-4b74-aac9-01f8f25d55a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.468044] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Downloading image file data 119dd2b4-b6de-42d5-adba-db818e4c2cd7 to the data store datastore1 {{(pid=62952) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 670.514569] env[62952]: DEBUG oslo_vmware.rw_handles [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/0448dc92-cb44-4a57-b5be-e39566a84600/119dd2b4-b6de-42d5-adba-db818e4c2cd7/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62952) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 670.573914] env[62952]: DEBUG nova.scheduler.client.report [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 670.582118] env[62952]: DEBUG nova.compute.manager [req-9eeb6ae9-f7cd-42e8-919d-7093a705b77a req-0a485ee2-aff0-4537-a888-ccaefe88657d service nova] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Received event network-changed-1d9877c9-1496-4654-a090-78b82fabb64e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 670.582244] env[62952]: DEBUG nova.compute.manager [req-9eeb6ae9-f7cd-42e8-919d-7093a705b77a req-0a485ee2-aff0-4537-a888-ccaefe88657d service nova] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Refreshing instance network info cache due to event network-changed-1d9877c9-1496-4654-a090-78b82fabb64e. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 670.582444] env[62952]: DEBUG oslo_concurrency.lockutils [req-9eeb6ae9-f7cd-42e8-919d-7093a705b77a req-0a485ee2-aff0-4537-a888-ccaefe88657d service nova] Acquiring lock "refresh_cache-0fd60040-c3ab-491f-a53e-f3297657367e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.582578] env[62952]: DEBUG oslo_concurrency.lockutils [req-9eeb6ae9-f7cd-42e8-919d-7093a705b77a req-0a485ee2-aff0-4537-a888-ccaefe88657d service nova] Acquired lock "refresh_cache-0fd60040-c3ab-491f-a53e-f3297657367e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.582733] env[62952]: DEBUG nova.network.neutron [req-9eeb6ae9-f7cd-42e8-919d-7093a705b77a req-0a485ee2-aff0-4537-a888-ccaefe88657d service nova] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Refreshing network info cache for port 1d9877c9-1496-4654-a090-78b82fabb64e {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 670.815503] env[62952]: ERROR nova.compute.manager [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1d9877c9-1496-4654-a090-78b82fabb64e, please check neutron logs for more information. [ 670.815503] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 670.815503] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.815503] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 670.815503] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.815503] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 670.815503] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.815503] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 670.815503] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.815503] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 670.815503] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.815503] env[62952]: ERROR nova.compute.manager raise self.value [ 670.815503] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.815503] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 670.815503] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.815503] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 670.816325] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.816325] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 670.816325] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1d9877c9-1496-4654-a090-78b82fabb64e, please check neutron logs for more information. [ 670.816325] env[62952]: ERROR nova.compute.manager [ 670.816325] env[62952]: Traceback (most recent call last): [ 670.816325] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 670.816325] env[62952]: listener.cb(fileno) [ 670.816325] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.816325] env[62952]: result = function(*args, **kwargs) [ 670.816325] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.816325] env[62952]: return func(*args, **kwargs) [ 670.816325] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.816325] env[62952]: raise e [ 670.816325] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.816325] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 670.816325] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.816325] env[62952]: created_port_ids = self._update_ports_for_instance( [ 670.816325] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.816325] env[62952]: with excutils.save_and_reraise_exception(): [ 670.816325] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.816325] env[62952]: self.force_reraise() [ 670.816325] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.816325] env[62952]: raise self.value [ 670.816325] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.816325] env[62952]: updated_port = self._update_port( [ 670.816325] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.816325] env[62952]: _ensure_no_port_binding_failure(port) [ 670.816325] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.816325] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 670.817472] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 1d9877c9-1496-4654-a090-78b82fabb64e, please check neutron logs for more information. [ 670.817472] env[62952]: Removing descriptor: 15 [ 671.082582] env[62952]: DEBUG oslo_concurrency.lockutils [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.996s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.083350] env[62952]: ERROR nova.compute.manager [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a5b1d0ee-1acc-4231-9031-c18558472a71, please check neutron logs for more information. [ 671.083350] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Traceback (most recent call last): [ 671.083350] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 671.083350] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] self.driver.spawn(context, instance, image_meta, [ 671.083350] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 671.083350] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 671.083350] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 671.083350] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] vm_ref = self.build_virtual_machine(instance, [ 671.083350] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 671.083350] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] vif_infos = vmwarevif.get_vif_info(self._session, [ 671.083350] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 671.083763] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] for vif in network_info: [ 671.083763] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 671.083763] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] return self._sync_wrapper(fn, *args, **kwargs) [ 671.083763] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 671.083763] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] self.wait() [ 671.083763] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 671.083763] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] self[:] = self._gt.wait() [ 671.083763] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 671.083763] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] return self._exit_event.wait() [ 671.083763] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 671.083763] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] result = hub.switch() [ 671.083763] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 671.083763] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] return self.greenlet.switch() [ 671.084217] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.084217] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] result = function(*args, **kwargs) [ 671.084217] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 671.084217] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] return func(*args, **kwargs) [ 671.084217] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 671.084217] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] raise e [ 671.084217] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.084217] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] nwinfo = self.network_api.allocate_for_instance( [ 671.084217] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 671.084217] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] created_port_ids = self._update_ports_for_instance( [ 671.084217] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 671.084217] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] with excutils.save_and_reraise_exception(): [ 671.084217] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.084680] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] self.force_reraise() [ 671.084680] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.084680] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] raise self.value [ 671.084680] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 671.084680] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] updated_port = self._update_port( [ 671.084680] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.084680] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] _ensure_no_port_binding_failure(port) [ 671.084680] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.084680] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] raise exception.PortBindingFailed(port_id=port['id']) [ 671.084680] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] nova.exception.PortBindingFailed: Binding failed for port a5b1d0ee-1acc-4231-9031-c18558472a71, please check neutron logs for more information. [ 671.084680] env[62952]: ERROR nova.compute.manager [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] [ 671.085068] env[62952]: DEBUG nova.compute.utils [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Binding failed for port a5b1d0ee-1acc-4231-9031-c18558472a71, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 671.085774] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.774s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.092826] env[62952]: DEBUG nova.compute.manager [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Build of instance 6c87dd22-6c47-461a-8129-585d1cd72a2a was re-scheduled: Binding failed for port a5b1d0ee-1acc-4231-9031-c18558472a71, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 671.093427] env[62952]: DEBUG nova.compute.manager [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 671.093645] env[62952]: DEBUG oslo_concurrency.lockutils [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Acquiring lock "refresh_cache-6c87dd22-6c47-461a-8129-585d1cd72a2a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.093952] env[62952]: DEBUG oslo_concurrency.lockutils [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Acquired lock "refresh_cache-6c87dd22-6c47-461a-8129-585d1cd72a2a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.094203] env[62952]: DEBUG nova.network.neutron [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 671.110611] env[62952]: DEBUG nova.compute.manager [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 671.113467] env[62952]: DEBUG nova.network.neutron [req-9eeb6ae9-f7cd-42e8-919d-7093a705b77a req-0a485ee2-aff0-4537-a888-ccaefe88657d service nova] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.139617] env[62952]: DEBUG nova.virt.hardware [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 671.139875] env[62952]: DEBUG nova.virt.hardware [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 671.140041] env[62952]: DEBUG nova.virt.hardware [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 671.140226] env[62952]: DEBUG nova.virt.hardware [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 671.140367] env[62952]: DEBUG nova.virt.hardware [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 671.140510] env[62952]: DEBUG nova.virt.hardware [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 671.140717] env[62952]: DEBUG nova.virt.hardware [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 671.140876] env[62952]: DEBUG nova.virt.hardware [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 671.141050] env[62952]: DEBUG nova.virt.hardware [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 671.141212] env[62952]: DEBUG nova.virt.hardware [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 671.141384] env[62952]: DEBUG nova.virt.hardware [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 671.142458] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b9d1f63-5460-4ed1-b6d1-cd1c87e99c30 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.150488] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f728064-045c-4c16-9785-5a9788665087 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.165477] env[62952]: ERROR nova.compute.manager [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1d9877c9-1496-4654-a090-78b82fabb64e, please check neutron logs for more information. [ 671.165477] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Traceback (most recent call last): [ 671.165477] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 671.165477] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] yield resources [ 671.165477] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 671.165477] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] self.driver.spawn(context, instance, image_meta, [ 671.165477] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 671.165477] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 671.165477] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 671.165477] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] vm_ref = self.build_virtual_machine(instance, [ 671.165477] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 671.165920] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] vif_infos = vmwarevif.get_vif_info(self._session, [ 671.165920] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 671.165920] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] for vif in network_info: [ 671.165920] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 671.165920] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] return self._sync_wrapper(fn, *args, **kwargs) [ 671.165920] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 671.165920] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] self.wait() [ 671.165920] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 671.165920] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] self[:] = self._gt.wait() [ 671.165920] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 671.165920] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] return self._exit_event.wait() [ 671.165920] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 671.165920] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] current.throw(*self._exc) [ 671.166376] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.166376] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] result = function(*args, **kwargs) [ 671.166376] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 671.166376] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] return func(*args, **kwargs) [ 671.166376] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 671.166376] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] raise e [ 671.166376] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.166376] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] nwinfo = self.network_api.allocate_for_instance( [ 671.166376] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 671.166376] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] created_port_ids = self._update_ports_for_instance( [ 671.166376] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 671.166376] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] with excutils.save_and_reraise_exception(): [ 671.166376] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.167067] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] self.force_reraise() [ 671.167067] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.167067] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] raise self.value [ 671.167067] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 671.167067] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] updated_port = self._update_port( [ 671.167067] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.167067] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] _ensure_no_port_binding_failure(port) [ 671.167067] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.167067] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] raise exception.PortBindingFailed(port_id=port['id']) [ 671.167067] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] nova.exception.PortBindingFailed: Binding failed for port 1d9877c9-1496-4654-a090-78b82fabb64e, please check neutron logs for more information. [ 671.167067] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] [ 671.167067] env[62952]: INFO nova.compute.manager [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Terminating instance [ 671.168639] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Acquiring lock "refresh_cache-0fd60040-c3ab-491f-a53e-f3297657367e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.211851] env[62952]: DEBUG nova.network.neutron [req-9eeb6ae9-f7cd-42e8-919d-7093a705b77a req-0a485ee2-aff0-4537-a888-ccaefe88657d service nova] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.224918] env[62952]: DEBUG oslo_vmware.rw_handles [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Completed reading data from the image iterator. {{(pid=62952) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 671.225127] env[62952]: DEBUG oslo_vmware.rw_handles [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/0448dc92-cb44-4a57-b5be-e39566a84600/119dd2b4-b6de-42d5-adba-db818e4c2cd7/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 671.611665] env[62952]: DEBUG nova.network.neutron [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.660408] env[62952]: DEBUG nova.network.neutron [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.715501] env[62952]: DEBUG oslo_concurrency.lockutils [req-9eeb6ae9-f7cd-42e8-919d-7093a705b77a req-0a485ee2-aff0-4537-a888-ccaefe88657d service nova] Releasing lock "refresh_cache-0fd60040-c3ab-491f-a53e-f3297657367e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.715909] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Acquired lock "refresh_cache-0fd60040-c3ab-491f-a53e-f3297657367e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.716109] env[62952]: DEBUG nova.network.neutron [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 671.941212] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5079a2b-ff4e-4d2d-93fc-f412f93fd875 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.948807] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc096aa-5744-431a-b062-bc33a672c78c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.978973] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12cb0665-2be5-487e-8ac5-34de7a4a437b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.986389] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de10a3b3-5901-4ffb-b672-eac74e0a2010 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.999477] env[62952]: DEBUG nova.compute.provider_tree [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.164053] env[62952]: DEBUG oslo_concurrency.lockutils [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Releasing lock "refresh_cache-6c87dd22-6c47-461a-8129-585d1cd72a2a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.164053] env[62952]: DEBUG nova.compute.manager [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 672.164053] env[62952]: DEBUG nova.compute.manager [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.164053] env[62952]: DEBUG nova.network.neutron [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 672.178974] env[62952]: DEBUG nova.network.neutron [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.235217] env[62952]: DEBUG nova.network.neutron [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.308685] env[62952]: DEBUG nova.network.neutron [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.502862] env[62952]: DEBUG nova.scheduler.client.report [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.593432] env[62952]: DEBUG nova.compute.manager [req-a93e4bb2-d12f-4c56-9691-2b0b86a05731 req-9ff730de-d7af-49c2-a48c-91bcf8013e43 service nova] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Received event network-vif-deleted-1d9877c9-1496-4654-a090-78b82fabb64e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 672.681547] env[62952]: DEBUG nova.network.neutron [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.813275] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Releasing lock "refresh_cache-0fd60040-c3ab-491f-a53e-f3297657367e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.813275] env[62952]: DEBUG nova.compute.manager [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 672.813553] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 672.813870] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0fd2e943-5b5f-4aaa-8016-9dc6076d1e79 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.822564] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728138b2-0259-4781-bb7a-1bc8dd91211d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.846410] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0fd60040-c3ab-491f-a53e-f3297657367e could not be found. [ 672.846725] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 672.846983] env[62952]: INFO nova.compute.manager [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 672.847470] env[62952]: DEBUG oslo.service.loopingcall [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 672.847470] env[62952]: DEBUG nova.compute.manager [-] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.847575] env[62952]: DEBUG nova.network.neutron [-] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 672.863182] env[62952]: DEBUG nova.network.neutron [-] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 673.008009] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.922s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.008718] env[62952]: ERROR nova.compute.manager [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0821cf2b-e08b-4f53-a4b6-da998fa92890, please check neutron logs for more information. [ 673.008718] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Traceback (most recent call last): [ 673.008718] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 673.008718] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] self.driver.spawn(context, instance, image_meta, [ 673.008718] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 673.008718] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 673.008718] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 673.008718] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] vm_ref = self.build_virtual_machine(instance, [ 673.008718] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 673.008718] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] vif_infos = vmwarevif.get_vif_info(self._session, [ 673.008718] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 673.009114] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] for vif in network_info: [ 673.009114] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 673.009114] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] return self._sync_wrapper(fn, *args, **kwargs) [ 673.009114] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 673.009114] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] self.wait() [ 673.009114] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 673.009114] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] self[:] = self._gt.wait() [ 673.009114] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 673.009114] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] return self._exit_event.wait() [ 673.009114] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 673.009114] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] result = hub.switch() [ 673.009114] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 673.009114] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] return self.greenlet.switch() [ 673.009535] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 673.009535] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] result = function(*args, **kwargs) [ 673.009535] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 673.009535] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] return func(*args, **kwargs) [ 673.009535] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 673.009535] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] raise e [ 673.009535] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 673.009535] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] nwinfo = self.network_api.allocate_for_instance( [ 673.009535] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 673.009535] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] created_port_ids = self._update_ports_for_instance( [ 673.009535] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 673.009535] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] with excutils.save_and_reraise_exception(): [ 673.009535] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 673.010020] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] self.force_reraise() [ 673.010020] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 673.010020] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] raise self.value [ 673.010020] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 673.010020] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] updated_port = self._update_port( [ 673.010020] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 673.010020] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] _ensure_no_port_binding_failure(port) [ 673.010020] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 673.010020] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] raise exception.PortBindingFailed(port_id=port['id']) [ 673.010020] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] nova.exception.PortBindingFailed: Binding failed for port 0821cf2b-e08b-4f53-a4b6-da998fa92890, please check neutron logs for more information. [ 673.010020] env[62952]: ERROR nova.compute.manager [instance: c377d16d-2641-483f-892b-42f6f067e9bc] [ 673.010422] env[62952]: DEBUG nova.compute.utils [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Binding failed for port 0821cf2b-e08b-4f53-a4b6-da998fa92890, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 673.010739] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.969s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.012154] env[62952]: INFO nova.compute.claims [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 673.015000] env[62952]: DEBUG nova.compute.manager [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Build of instance c377d16d-2641-483f-892b-42f6f067e9bc was re-scheduled: Binding failed for port 0821cf2b-e08b-4f53-a4b6-da998fa92890, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 673.015161] env[62952]: DEBUG nova.compute.manager [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 673.015374] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Acquiring lock "refresh_cache-c377d16d-2641-483f-892b-42f6f067e9bc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.015513] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Acquired lock "refresh_cache-c377d16d-2641-483f-892b-42f6f067e9bc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.015675] env[62952]: DEBUG nova.network.neutron [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 673.184025] env[62952]: INFO nova.compute.manager [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 6c87dd22-6c47-461a-8129-585d1cd72a2a] Took 1.02 seconds to deallocate network for instance. [ 673.365553] env[62952]: DEBUG nova.network.neutron [-] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.531931] env[62952]: DEBUG nova.network.neutron [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 673.703998] env[62952]: DEBUG nova.network.neutron [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.828025] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Downloaded image file data 119dd2b4-b6de-42d5-adba-db818e4c2cd7 to vmware_temp/0448dc92-cb44-4a57-b5be-e39566a84600/119dd2b4-b6de-42d5-adba-db818e4c2cd7/tmp-sparse.vmdk on the data store datastore1 {{(pid=62952) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 673.829755] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Caching image {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 673.830015] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Copying Virtual Disk [datastore1] vmware_temp/0448dc92-cb44-4a57-b5be-e39566a84600/119dd2b4-b6de-42d5-adba-db818e4c2cd7/tmp-sparse.vmdk to [datastore1] vmware_temp/0448dc92-cb44-4a57-b5be-e39566a84600/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 673.830695] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-77f5ec29-74df-4ba5-ab8e-202025c5bfce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.838421] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Waiting for the task: (returnval){ [ 673.838421] env[62952]: value = "task-1263053" [ 673.838421] env[62952]: _type = "Task" [ 673.838421] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.847351] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263053, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.868339] env[62952]: INFO nova.compute.manager [-] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Took 1.02 seconds to deallocate network for instance. [ 673.870662] env[62952]: DEBUG nova.compute.claims [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 673.870874] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.206900] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Releasing lock "refresh_cache-c377d16d-2641-483f-892b-42f6f067e9bc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.207496] env[62952]: DEBUG nova.compute.manager [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 674.207496] env[62952]: DEBUG nova.compute.manager [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 674.207496] env[62952]: DEBUG nova.network.neutron [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 674.210250] env[62952]: INFO nova.scheduler.client.report [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Deleted allocations for instance 6c87dd22-6c47-461a-8129-585d1cd72a2a [ 674.236095] env[62952]: DEBUG nova.network.neutron [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 674.350746] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263053, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.446088] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2d17ae0-a249-40d1-b5ab-a772d067f634 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.454770] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0348ba2-1691-4f20-ae19-ffa89fa08ebc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.488556] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3766e30-5bbb-454e-a9a2-5e08bdbd7c8a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.495966] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ca3159-8a81-42d6-9fe5-5a5860092fe9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.509158] env[62952]: DEBUG nova.compute.provider_tree [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.720794] env[62952]: DEBUG oslo_concurrency.lockutils [None req-068b7abc-1000-4268-b6ca-f1750773e261 tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Lock "6c87dd22-6c47-461a-8129-585d1cd72a2a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.850s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.738060] env[62952]: DEBUG nova.network.neutron [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.849678] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263053, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.660138} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.849945] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Copied Virtual Disk [datastore1] vmware_temp/0448dc92-cb44-4a57-b5be-e39566a84600/119dd2b4-b6de-42d5-adba-db818e4c2cd7/tmp-sparse.vmdk to [datastore1] vmware_temp/0448dc92-cb44-4a57-b5be-e39566a84600/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 674.850141] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Deleting the datastore file [datastore1] vmware_temp/0448dc92-cb44-4a57-b5be-e39566a84600/119dd2b4-b6de-42d5-adba-db818e4c2cd7/tmp-sparse.vmdk {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 674.850387] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8eb94dac-2cde-4345-95fe-b72305a457ba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.857196] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Waiting for the task: (returnval){ [ 674.857196] env[62952]: value = "task-1263054" [ 674.857196] env[62952]: _type = "Task" [ 674.857196] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.865594] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263054, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.011762] env[62952]: DEBUG nova.scheduler.client.report [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.223416] env[62952]: DEBUG nova.compute.manager [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 675.241054] env[62952]: INFO nova.compute.manager [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] [instance: c377d16d-2641-483f-892b-42f6f067e9bc] Took 1.03 seconds to deallocate network for instance. [ 675.366781] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263054, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023919} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.367084] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 675.367301] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Moving file from [datastore1] vmware_temp/0448dc92-cb44-4a57-b5be-e39566a84600/119dd2b4-b6de-42d5-adba-db818e4c2cd7 to [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7. {{(pid=62952) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 675.367549] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-2de2f94e-9bf9-4606-afb5-4f1ac31e7825 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.373862] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Waiting for the task: (returnval){ [ 675.373862] env[62952]: value = "task-1263055" [ 675.373862] env[62952]: _type = "Task" [ 675.373862] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.381902] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263055, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.516476] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.506s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.517081] env[62952]: DEBUG nova.compute.manager [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 675.521019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.302s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.744230] env[62952]: DEBUG oslo_concurrency.lockutils [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.884945] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263055, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.029257} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.885223] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] File moved {{(pid=62952) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 675.885407] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Cleaning up location [datastore1] vmware_temp/0448dc92-cb44-4a57-b5be-e39566a84600 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 675.885566] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Deleting the datastore file [datastore1] vmware_temp/0448dc92-cb44-4a57-b5be-e39566a84600 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 675.885812] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de0a9d26-fde2-4fab-a628-084248fc8185 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.892468] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Waiting for the task: (returnval){ [ 675.892468] env[62952]: value = "task-1263056" [ 675.892468] env[62952]: _type = "Task" [ 675.892468] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.900102] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263056, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.025364] env[62952]: DEBUG nova.compute.utils [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 676.029999] env[62952]: DEBUG nova.compute.manager [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 676.030207] env[62952]: DEBUG nova.network.neutron [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 676.075456] env[62952]: DEBUG nova.policy [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ef72aa6e64d4378883e53894dfae086', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e09d6d5c5fe8472b88bf4e7c3d0d0467', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 676.159473] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Acquiring lock "9979dd1d-9e99-47cf-a81a-fd7d89531f51" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.159705] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Lock "9979dd1d-9e99-47cf-a81a-fd7d89531f51" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.273470] env[62952]: INFO nova.scheduler.client.report [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Deleted allocations for instance c377d16d-2641-483f-892b-42f6f067e9bc [ 676.354685] env[62952]: DEBUG nova.network.neutron [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Successfully created port: c97da0ad-3a77-4848-ac45-978faf4e8ce6 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 676.401688] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263056, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.02462} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.402122] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 676.405176] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51a0bc4f-2ac5-4ed5-a571-548ada87960e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.413244] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Waiting for the task: (returnval){ [ 676.413244] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524781e9-e967-f2bd-264a-8cc204a68b91" [ 676.413244] env[62952]: _type = "Task" [ 676.413244] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.420164] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524781e9-e967-f2bd-264a-8cc204a68b91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.422052] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da731898-0bf8-485a-86e7-202f45e1c07b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.428932] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e92c773-609e-4510-aa1d-6a59f6452e8f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.458590] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4abddb38-c4db-481c-8a9f-061cb2445126 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.465723] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c505ea-b6bc-48b7-afea-a49a498c60bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.480111] env[62952]: DEBUG nova.compute.provider_tree [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.531196] env[62952]: DEBUG nova.compute.manager [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 676.784887] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a20b5de-df44-4862-b2bc-320aa7386c87 tempest-ServersTestFqdnHostnames-1768710099 tempest-ServersTestFqdnHostnames-1768710099-project-member] Lock "c377d16d-2641-483f-892b-42f6f067e9bc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.913s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.923340] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524781e9-e967-f2bd-264a-8cc204a68b91, 'name': SearchDatastore_Task, 'duration_secs': 0.008363} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.923340] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.923340] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] dc025711-45be-49b6-a765-f7687b94e6a1/dc025711-45be-49b6-a765-f7687b94e6a1.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 676.923340] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-35a49e96-9566-4539-b514-c1f560e863d9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.928955] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Waiting for the task: (returnval){ [ 676.928955] env[62952]: value = "task-1263057" [ 676.928955] env[62952]: _type = "Task" [ 676.928955] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.936492] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263057, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.983919] env[62952]: DEBUG nova.scheduler.client.report [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.024939] env[62952]: DEBUG nova.compute.manager [req-a0ff1e45-a624-4dbb-a585-672ae27dd4e5 req-4447436b-f7ab-4e31-8d78-3b4986245c8f service nova] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Received event network-changed-c97da0ad-3a77-4848-ac45-978faf4e8ce6 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 677.025226] env[62952]: DEBUG nova.compute.manager [req-a0ff1e45-a624-4dbb-a585-672ae27dd4e5 req-4447436b-f7ab-4e31-8d78-3b4986245c8f service nova] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Refreshing instance network info cache due to event network-changed-c97da0ad-3a77-4848-ac45-978faf4e8ce6. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 677.025448] env[62952]: DEBUG oslo_concurrency.lockutils [req-a0ff1e45-a624-4dbb-a585-672ae27dd4e5 req-4447436b-f7ab-4e31-8d78-3b4986245c8f service nova] Acquiring lock "refresh_cache-4c6f81ba-85f5-4ad8-a3c8-57867c613baa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.025585] env[62952]: DEBUG oslo_concurrency.lockutils [req-a0ff1e45-a624-4dbb-a585-672ae27dd4e5 req-4447436b-f7ab-4e31-8d78-3b4986245c8f service nova] Acquired lock "refresh_cache-4c6f81ba-85f5-4ad8-a3c8-57867c613baa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.025741] env[62952]: DEBUG nova.network.neutron [req-a0ff1e45-a624-4dbb-a585-672ae27dd4e5 req-4447436b-f7ab-4e31-8d78-3b4986245c8f service nova] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Refreshing network info cache for port c97da0ad-3a77-4848-ac45-978faf4e8ce6 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 677.218868] env[62952]: ERROR nova.compute.manager [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c97da0ad-3a77-4848-ac45-978faf4e8ce6, please check neutron logs for more information. [ 677.218868] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 677.218868] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.218868] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 677.218868] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.218868] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 677.218868] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.218868] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 677.218868] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.218868] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 677.218868] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.218868] env[62952]: ERROR nova.compute.manager raise self.value [ 677.218868] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.218868] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 677.218868] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.218868] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 677.219389] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.219389] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 677.219389] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c97da0ad-3a77-4848-ac45-978faf4e8ce6, please check neutron logs for more information. [ 677.219389] env[62952]: ERROR nova.compute.manager [ 677.219389] env[62952]: Traceback (most recent call last): [ 677.219389] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 677.219389] env[62952]: listener.cb(fileno) [ 677.219389] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.219389] env[62952]: result = function(*args, **kwargs) [ 677.219389] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.219389] env[62952]: return func(*args, **kwargs) [ 677.219389] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.219389] env[62952]: raise e [ 677.219389] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.219389] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 677.219389] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.219389] env[62952]: created_port_ids = self._update_ports_for_instance( [ 677.219389] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.219389] env[62952]: with excutils.save_and_reraise_exception(): [ 677.219389] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.219389] env[62952]: self.force_reraise() [ 677.219389] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.219389] env[62952]: raise self.value [ 677.219389] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.219389] env[62952]: updated_port = self._update_port( [ 677.219389] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.219389] env[62952]: _ensure_no_port_binding_failure(port) [ 677.219389] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.219389] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 677.220059] env[62952]: nova.exception.PortBindingFailed: Binding failed for port c97da0ad-3a77-4848-ac45-978faf4e8ce6, please check neutron logs for more information. [ 677.220059] env[62952]: Removing descriptor: 15 [ 677.288925] env[62952]: DEBUG nova.compute.manager [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 677.438995] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263057, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.490076] env[62952]: DEBUG oslo_concurrency.lockutils [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.969s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.490787] env[62952]: ERROR nova.compute.manager [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 270db801-02b2-419b-9df9-eb2c926b8e24, please check neutron logs for more information. [ 677.490787] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Traceback (most recent call last): [ 677.490787] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.490787] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] self.driver.spawn(context, instance, image_meta, [ 677.490787] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 677.490787] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.490787] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.490787] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] vm_ref = self.build_virtual_machine(instance, [ 677.490787] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.490787] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.490787] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.491309] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] for vif in network_info: [ 677.491309] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.491309] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] return self._sync_wrapper(fn, *args, **kwargs) [ 677.491309] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.491309] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] self.wait() [ 677.491309] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.491309] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] self[:] = self._gt.wait() [ 677.491309] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.491309] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] return self._exit_event.wait() [ 677.491309] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 677.491309] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] result = hub.switch() [ 677.491309] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 677.491309] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] return self.greenlet.switch() [ 677.491663] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.491663] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] result = function(*args, **kwargs) [ 677.491663] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.491663] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] return func(*args, **kwargs) [ 677.491663] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.491663] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] raise e [ 677.491663] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.491663] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] nwinfo = self.network_api.allocate_for_instance( [ 677.491663] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.491663] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] created_port_ids = self._update_ports_for_instance( [ 677.491663] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.491663] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] with excutils.save_and_reraise_exception(): [ 677.491663] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.491963] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] self.force_reraise() [ 677.491963] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.491963] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] raise self.value [ 677.491963] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.491963] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] updated_port = self._update_port( [ 677.491963] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.491963] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] _ensure_no_port_binding_failure(port) [ 677.491963] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.491963] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] raise exception.PortBindingFailed(port_id=port['id']) [ 677.491963] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] nova.exception.PortBindingFailed: Binding failed for port 270db801-02b2-419b-9df9-eb2c926b8e24, please check neutron logs for more information. [ 677.491963] env[62952]: ERROR nova.compute.manager [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] [ 677.492245] env[62952]: DEBUG nova.compute.utils [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Binding failed for port 270db801-02b2-419b-9df9-eb2c926b8e24, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 677.492722] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.727s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.495758] env[62952]: DEBUG nova.compute.manager [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Build of instance 25e10d02-8956-4d53-a1a4-b977da38f1c8 was re-scheduled: Binding failed for port 270db801-02b2-419b-9df9-eb2c926b8e24, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 677.496760] env[62952]: DEBUG nova.compute.manager [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 677.496760] env[62952]: DEBUG oslo_concurrency.lockutils [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Acquiring lock "refresh_cache-25e10d02-8956-4d53-a1a4-b977da38f1c8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.496760] env[62952]: DEBUG oslo_concurrency.lockutils [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Acquired lock "refresh_cache-25e10d02-8956-4d53-a1a4-b977da38f1c8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.496894] env[62952]: DEBUG nova.network.neutron [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 677.539857] env[62952]: DEBUG nova.compute.manager [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 677.545522] env[62952]: DEBUG nova.network.neutron [req-a0ff1e45-a624-4dbb-a585-672ae27dd4e5 req-4447436b-f7ab-4e31-8d78-3b4986245c8f service nova] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.564335] env[62952]: DEBUG nova.virt.hardware [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 677.564471] env[62952]: DEBUG nova.virt.hardware [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 677.564750] env[62952]: DEBUG nova.virt.hardware [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 677.564813] env[62952]: DEBUG nova.virt.hardware [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 677.564943] env[62952]: DEBUG nova.virt.hardware [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 677.565096] env[62952]: DEBUG nova.virt.hardware [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 677.565306] env[62952]: DEBUG nova.virt.hardware [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 677.565460] env[62952]: DEBUG nova.virt.hardware [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 677.565618] env[62952]: DEBUG nova.virt.hardware [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 677.565776] env[62952]: DEBUG nova.virt.hardware [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 677.565953] env[62952]: DEBUG nova.virt.hardware [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.566804] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de0f72cb-862c-4c6f-b3cd-227dade6d050 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.574484] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290f4583-d575-4953-8630-b78ffeaa7631 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.590431] env[62952]: ERROR nova.compute.manager [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c97da0ad-3a77-4848-ac45-978faf4e8ce6, please check neutron logs for more information. [ 677.590431] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Traceback (most recent call last): [ 677.590431] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 677.590431] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] yield resources [ 677.590431] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.590431] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] self.driver.spawn(context, instance, image_meta, [ 677.590431] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 677.590431] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.590431] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.590431] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] vm_ref = self.build_virtual_machine(instance, [ 677.590431] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.590748] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.590748] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.590748] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] for vif in network_info: [ 677.590748] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.590748] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] return self._sync_wrapper(fn, *args, **kwargs) [ 677.590748] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.590748] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] self.wait() [ 677.590748] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.590748] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] self[:] = self._gt.wait() [ 677.590748] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.590748] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] return self._exit_event.wait() [ 677.590748] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 677.590748] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] current.throw(*self._exc) [ 677.591088] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.591088] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] result = function(*args, **kwargs) [ 677.591088] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.591088] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] return func(*args, **kwargs) [ 677.591088] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.591088] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] raise e [ 677.591088] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.591088] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] nwinfo = self.network_api.allocate_for_instance( [ 677.591088] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.591088] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] created_port_ids = self._update_ports_for_instance( [ 677.591088] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.591088] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] with excutils.save_and_reraise_exception(): [ 677.591088] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.591439] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] self.force_reraise() [ 677.591439] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.591439] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] raise self.value [ 677.591439] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.591439] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] updated_port = self._update_port( [ 677.591439] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.591439] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] _ensure_no_port_binding_failure(port) [ 677.591439] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.591439] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] raise exception.PortBindingFailed(port_id=port['id']) [ 677.591439] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] nova.exception.PortBindingFailed: Binding failed for port c97da0ad-3a77-4848-ac45-978faf4e8ce6, please check neutron logs for more information. [ 677.591439] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] [ 677.591439] env[62952]: INFO nova.compute.manager [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Terminating instance [ 677.594878] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Acquiring lock "refresh_cache-4c6f81ba-85f5-4ad8-a3c8-57867c613baa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.629775] env[62952]: DEBUG nova.network.neutron [req-a0ff1e45-a624-4dbb-a585-672ae27dd4e5 req-4447436b-f7ab-4e31-8d78-3b4986245c8f service nova] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.810215] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.941370] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263057, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.028584] env[62952]: DEBUG nova.network.neutron [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.105773] env[62952]: DEBUG nova.network.neutron [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.132682] env[62952]: DEBUG oslo_concurrency.lockutils [req-a0ff1e45-a624-4dbb-a585-672ae27dd4e5 req-4447436b-f7ab-4e31-8d78-3b4986245c8f service nova] Releasing lock "refresh_cache-4c6f81ba-85f5-4ad8-a3c8-57867c613baa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.133077] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Acquired lock "refresh_cache-4c6f81ba-85f5-4ad8-a3c8-57867c613baa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.134017] env[62952]: DEBUG nova.network.neutron [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 678.365423] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c181f7d-c1b6-402b-9713-a2e15e134391 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.374366] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76c0273e-26c1-47e2-a3d6-c7cf016b91e3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.406308] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08cfa59a-5b8d-457e-805b-70d89ad55bcb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.413360] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-486ef4d9-24aa-4c95-827b-467355a0ee1c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.429762] env[62952]: DEBUG nova.compute.provider_tree [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.438999] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263057, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.608379] env[62952]: DEBUG oslo_concurrency.lockutils [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Releasing lock "refresh_cache-25e10d02-8956-4d53-a1a4-b977da38f1c8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.608638] env[62952]: DEBUG nova.compute.manager [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 678.608805] env[62952]: DEBUG nova.compute.manager [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.608972] env[62952]: DEBUG nova.network.neutron [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 678.628254] env[62952]: DEBUG nova.network.neutron [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.653391] env[62952]: DEBUG nova.network.neutron [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.741904] env[62952]: DEBUG nova.network.neutron [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.935898] env[62952]: DEBUG nova.scheduler.client.report [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 678.951733] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263057, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.566412} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.951959] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] dc025711-45be-49b6-a765-f7687b94e6a1/dc025711-45be-49b6-a765-f7687b94e6a1.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 678.952515] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 678.952515] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-755bd5d7-1bfb-4c5a-9a3b-955b04f5b72b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.959557] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Waiting for the task: (returnval){ [ 678.959557] env[62952]: value = "task-1263058" [ 678.959557] env[62952]: _type = "Task" [ 678.959557] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.968924] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263058, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.078087] env[62952]: DEBUG nova.compute.manager [req-fbf5ee90-df72-46a6-9f1f-07601feb9837 req-d98ef2a0-0a1e-48c5-9673-f0d475d1a755 service nova] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Received event network-vif-deleted-c97da0ad-3a77-4848-ac45-978faf4e8ce6 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 679.129402] env[62952]: DEBUG nova.network.neutron [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.244575] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Releasing lock "refresh_cache-4c6f81ba-85f5-4ad8-a3c8-57867c613baa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.244679] env[62952]: DEBUG nova.compute.manager [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 679.245490] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 679.245490] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5d462978-a381-42a5-bf24-fad852dde5f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.254277] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb76cba-df20-4f36-b64b-7800e7729c9e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.275547] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4c6f81ba-85f5-4ad8-a3c8-57867c613baa could not be found. [ 679.275761] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 679.275938] env[62952]: INFO nova.compute.manager [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Took 0.03 seconds to destroy the instance on the hypervisor. [ 679.276185] env[62952]: DEBUG oslo.service.loopingcall [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 679.276400] env[62952]: DEBUG nova.compute.manager [-] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 679.276494] env[62952]: DEBUG nova.network.neutron [-] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 679.300933] env[62952]: DEBUG nova.network.neutron [-] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.452441] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.960s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.453489] env[62952]: ERROR nova.compute.manager [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e2f7d15e-3a04-4206-ab85-de44810e6577, please check neutron logs for more information. [ 679.453489] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Traceback (most recent call last): [ 679.453489] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.453489] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] self.driver.spawn(context, instance, image_meta, [ 679.453489] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 679.453489] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.453489] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.453489] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] vm_ref = self.build_virtual_machine(instance, [ 679.453489] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.453489] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.453489] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.453785] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] for vif in network_info: [ 679.453785] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.453785] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] return self._sync_wrapper(fn, *args, **kwargs) [ 679.453785] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.453785] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] self.wait() [ 679.453785] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.453785] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] self[:] = self._gt.wait() [ 679.453785] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.453785] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] return self._exit_event.wait() [ 679.453785] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 679.453785] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] current.throw(*self._exc) [ 679.453785] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.453785] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] result = function(*args, **kwargs) [ 679.454096] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.454096] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] return func(*args, **kwargs) [ 679.454096] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.454096] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] raise e [ 679.454096] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.454096] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] nwinfo = self.network_api.allocate_for_instance( [ 679.454096] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 679.454096] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] created_port_ids = self._update_ports_for_instance( [ 679.454096] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 679.454096] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] with excutils.save_and_reraise_exception(): [ 679.454096] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.454096] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] self.force_reraise() [ 679.454096] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.454420] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] raise self.value [ 679.454420] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 679.454420] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] updated_port = self._update_port( [ 679.454420] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.454420] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] _ensure_no_port_binding_failure(port) [ 679.454420] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.454420] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] raise exception.PortBindingFailed(port_id=port['id']) [ 679.454420] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] nova.exception.PortBindingFailed: Binding failed for port e2f7d15e-3a04-4206-ab85-de44810e6577, please check neutron logs for more information. [ 679.454420] env[62952]: ERROR nova.compute.manager [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] [ 679.454420] env[62952]: DEBUG nova.compute.utils [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Binding failed for port e2f7d15e-3a04-4206-ab85-de44810e6577, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 679.455520] env[62952]: DEBUG oslo_concurrency.lockutils [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.103s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.458950] env[62952]: DEBUG nova.compute.manager [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Build of instance 78a0acf9-12f5-4699-a4b8-2cbefcbd343b was re-scheduled: Binding failed for port e2f7d15e-3a04-4206-ab85-de44810e6577, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 679.459112] env[62952]: DEBUG nova.compute.manager [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 679.459292] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Acquiring lock "refresh_cache-78a0acf9-12f5-4699-a4b8-2cbefcbd343b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.459439] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Acquired lock "refresh_cache-78a0acf9-12f5-4699-a4b8-2cbefcbd343b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.459596] env[62952]: DEBUG nova.network.neutron [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 679.469080] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263058, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066953} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.469913] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 679.470680] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c420ee2-6a38-4623-93e1-fa6fca3fd1bd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.494445] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Reconfiguring VM instance instance-00000012 to attach disk [datastore1] dc025711-45be-49b6-a765-f7687b94e6a1/dc025711-45be-49b6-a765-f7687b94e6a1.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 679.495412] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c524ea27-ee2b-4cb4-b435-0aa0b0f19edb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.516693] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Waiting for the task: (returnval){ [ 679.516693] env[62952]: value = "task-1263059" [ 679.516693] env[62952]: _type = "Task" [ 679.516693] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.525652] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263059, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.632335] env[62952]: INFO nova.compute.manager [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] [instance: 25e10d02-8956-4d53-a1a4-b977da38f1c8] Took 1.02 seconds to deallocate network for instance. [ 679.803544] env[62952]: DEBUG nova.network.neutron [-] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.983740] env[62952]: DEBUG nova.network.neutron [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.027922] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263059, 'name': ReconfigVM_Task, 'duration_secs': 0.266219} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.027922] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Reconfigured VM instance instance-00000012 to attach disk [datastore1] dc025711-45be-49b6-a765-f7687b94e6a1/dc025711-45be-49b6-a765-f7687b94e6a1.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 680.027922] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-79af33f1-4b0a-40d7-acdd-a9ec2cb6a6d1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.041978] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Waiting for the task: (returnval){ [ 680.041978] env[62952]: value = "task-1263060" [ 680.041978] env[62952]: _type = "Task" [ 680.041978] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.050257] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263060, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.068775] env[62952]: DEBUG nova.network.neutron [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.306296] env[62952]: INFO nova.compute.manager [-] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Took 1.03 seconds to deallocate network for instance. [ 680.311645] env[62952]: DEBUG nova.compute.claims [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 680.311645] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.352824] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c02115-e933-4835-a9db-6153b1e901e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.360456] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6940f4c-bb36-43ed-8738-6de2ad15bafe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.392077] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46fcd877-1dcf-4583-8b30-c0f0cabe2215 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.399133] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b48eade5-6a33-4589-8f8e-6fa6c52e77ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.412237] env[62952]: DEBUG nova.compute.provider_tree [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.549137] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263060, 'name': Rename_Task, 'duration_secs': 0.132767} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.549404] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 680.549635] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d571a449-6ff4-4421-b037-be068261fa82 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.555299] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Waiting for the task: (returnval){ [ 680.555299] env[62952]: value = "task-1263061" [ 680.555299] env[62952]: _type = "Task" [ 680.555299] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.562467] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263061, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.571041] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Releasing lock "refresh_cache-78a0acf9-12f5-4699-a4b8-2cbefcbd343b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.571272] env[62952]: DEBUG nova.compute.manager [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 680.571457] env[62952]: DEBUG nova.compute.manager [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 680.571622] env[62952]: DEBUG nova.network.neutron [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 680.587644] env[62952]: DEBUG nova.network.neutron [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.679957] env[62952]: INFO nova.scheduler.client.report [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Deleted allocations for instance 25e10d02-8956-4d53-a1a4-b977da38f1c8 [ 680.918036] env[62952]: DEBUG nova.scheduler.client.report [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.067136] env[62952]: DEBUG oslo_vmware.api [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263061, 'name': PowerOnVM_Task, 'duration_secs': 0.413947} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.067367] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 681.067634] env[62952]: INFO nova.compute.manager [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Took 12.39 seconds to spawn the instance on the hypervisor. [ 681.067921] env[62952]: DEBUG nova.compute.manager [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 681.068668] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d15df129-0486-4981-8809-0e230c3eda47 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.093562] env[62952]: DEBUG nova.network.neutron [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.190294] env[62952]: DEBUG oslo_concurrency.lockutils [None req-536c991d-45b6-4609-b48f-9c00a7eac37a tempest-ServersAdminNegativeTestJSON-52600179 tempest-ServersAdminNegativeTestJSON-52600179-project-member] Lock "25e10d02-8956-4d53-a1a4-b977da38f1c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.437s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.422913] env[62952]: DEBUG oslo_concurrency.lockutils [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.967s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.423357] env[62952]: ERROR nova.compute.manager [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f3a6797c-20d9-4a73-b4ef-f147a2a28c20, please check neutron logs for more information. [ 681.423357] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Traceback (most recent call last): [ 681.423357] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 681.423357] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] self.driver.spawn(context, instance, image_meta, [ 681.423357] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 681.423357] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.423357] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.423357] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] vm_ref = self.build_virtual_machine(instance, [ 681.423357] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.423357] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.423357] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.423628] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] for vif in network_info: [ 681.423628] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 681.423628] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] return self._sync_wrapper(fn, *args, **kwargs) [ 681.423628] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 681.423628] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] self.wait() [ 681.423628] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 681.423628] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] self[:] = self._gt.wait() [ 681.423628] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.423628] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] return self._exit_event.wait() [ 681.423628] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 681.423628] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] current.throw(*self._exc) [ 681.423628] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.423628] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] result = function(*args, **kwargs) [ 681.423922] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 681.423922] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] return func(*args, **kwargs) [ 681.423922] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.423922] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] raise e [ 681.423922] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.423922] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] nwinfo = self.network_api.allocate_for_instance( [ 681.423922] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 681.423922] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] created_port_ids = self._update_ports_for_instance( [ 681.423922] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 681.423922] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] with excutils.save_and_reraise_exception(): [ 681.423922] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.423922] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] self.force_reraise() [ 681.423922] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.424209] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] raise self.value [ 681.424209] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 681.424209] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] updated_port = self._update_port( [ 681.424209] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.424209] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] _ensure_no_port_binding_failure(port) [ 681.424209] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.424209] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] raise exception.PortBindingFailed(port_id=port['id']) [ 681.424209] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] nova.exception.PortBindingFailed: Binding failed for port f3a6797c-20d9-4a73-b4ef-f147a2a28c20, please check neutron logs for more information. [ 681.424209] env[62952]: ERROR nova.compute.manager [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] [ 681.424209] env[62952]: DEBUG nova.compute.utils [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Binding failed for port f3a6797c-20d9-4a73-b4ef-f147a2a28c20, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 681.425378] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.634s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.428813] env[62952]: DEBUG nova.compute.manager [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Build of instance 4eb9ea32-b2eb-440c-871b-6942590d9c01 was re-scheduled: Binding failed for port f3a6797c-20d9-4a73-b4ef-f147a2a28c20, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 681.428870] env[62952]: DEBUG nova.compute.manager [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 681.429112] env[62952]: DEBUG oslo_concurrency.lockutils [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Acquiring lock "refresh_cache-4eb9ea32-b2eb-440c-871b-6942590d9c01" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.429255] env[62952]: DEBUG oslo_concurrency.lockutils [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Acquired lock "refresh_cache-4eb9ea32-b2eb-440c-871b-6942590d9c01" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.429413] env[62952]: DEBUG nova.network.neutron [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 681.586425] env[62952]: INFO nova.compute.manager [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Took 34.92 seconds to build instance. [ 681.600354] env[62952]: INFO nova.compute.manager [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] [instance: 78a0acf9-12f5-4699-a4b8-2cbefcbd343b] Took 1.03 seconds to deallocate network for instance. [ 681.692743] env[62952]: DEBUG nova.compute.manager [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 681.962209] env[62952]: DEBUG nova.network.neutron [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 682.067940] env[62952]: DEBUG nova.network.neutron [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.088650] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b0f99a9-e3d4-43af-a5ae-9f240a804e00 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Lock "dc025711-45be-49b6-a765-f7687b94e6a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.255s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.225083] env[62952]: DEBUG oslo_concurrency.lockutils [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.314733] env[62952]: DEBUG nova.compute.manager [None req-7062643b-666b-4a62-8457-6ee195941655 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 682.315665] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9240f23a-9c19-474e-967b-de1ea9f7d03b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.438146] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-291a4ea8-1804-4be3-8567-7464d6d1353d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.438146] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f12c3d7c-7030-4b56-9a02-3c3e3f5134dc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.459463] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46026ccf-9349-4cbb-818d-a62f003578d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.467028] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c4733c-b2ea-41d0-9800-44e04935ed48 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.480305] env[62952]: DEBUG nova.compute.provider_tree [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 682.535363] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Acquiring lock "dc025711-45be-49b6-a765-f7687b94e6a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.535630] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Lock "dc025711-45be-49b6-a765-f7687b94e6a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.535843] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Acquiring lock "dc025711-45be-49b6-a765-f7687b94e6a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.536028] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Lock "dc025711-45be-49b6-a765-f7687b94e6a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.540026] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Lock "dc025711-45be-49b6-a765-f7687b94e6a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.540026] env[62952]: INFO nova.compute.manager [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Terminating instance [ 682.541687] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Acquiring lock "refresh_cache-dc025711-45be-49b6-a765-f7687b94e6a1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.541832] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Acquired lock "refresh_cache-dc025711-45be-49b6-a765-f7687b94e6a1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.541993] env[62952]: DEBUG nova.network.neutron [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 682.571243] env[62952]: DEBUG oslo_concurrency.lockutils [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Releasing lock "refresh_cache-4eb9ea32-b2eb-440c-871b-6942590d9c01" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.571579] env[62952]: DEBUG nova.compute.manager [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 682.571819] env[62952]: DEBUG nova.compute.manager [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 682.572054] env[62952]: DEBUG nova.network.neutron [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 682.590336] env[62952]: DEBUG nova.network.neutron [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 682.606308] env[62952]: DEBUG nova.compute.manager [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 682.638502] env[62952]: INFO nova.scheduler.client.report [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Deleted allocations for instance 78a0acf9-12f5-4699-a4b8-2cbefcbd343b [ 682.835847] env[62952]: INFO nova.compute.manager [None req-7062643b-666b-4a62-8457-6ee195941655 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] instance snapshotting [ 682.836096] env[62952]: DEBUG nova.objects.instance [None req-7062643b-666b-4a62-8457-6ee195941655 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Lazy-loading 'flavor' on Instance uuid dc025711-45be-49b6-a765-f7687b94e6a1 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 682.983865] env[62952]: DEBUG nova.scheduler.client.report [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 683.073022] env[62952]: DEBUG nova.network.neutron [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 683.099083] env[62952]: DEBUG nova.network.neutron [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.121227] env[62952]: DEBUG nova.network.neutron [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.135861] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.152257] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f63d7c6-918e-46a7-80bb-6581f2eb4c23 tempest-VolumesAssistedSnapshotsTest-474997348 tempest-VolumesAssistedSnapshotsTest-474997348-project-member] Lock "78a0acf9-12f5-4699-a4b8-2cbefcbd343b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.732s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.343227] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e32e97-f33a-435a-92e1-3c05938d0b44 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.363808] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9b309ce-27a0-4bc9-8871-926162613191 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.489367] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.064s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.490098] env[62952]: ERROR nova.compute.manager [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e46c02d0-3a45-4099-8006-a529a1c1f2df, please check neutron logs for more information. [ 683.490098] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Traceback (most recent call last): [ 683.490098] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 683.490098] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] self.driver.spawn(context, instance, image_meta, [ 683.490098] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 683.490098] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] self._vmops.spawn(context, instance, image_meta, injected_files, [ 683.490098] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 683.490098] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] vm_ref = self.build_virtual_machine(instance, [ 683.490098] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 683.490098] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] vif_infos = vmwarevif.get_vif_info(self._session, [ 683.490098] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 683.490391] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] for vif in network_info: [ 683.490391] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 683.490391] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] return self._sync_wrapper(fn, *args, **kwargs) [ 683.490391] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 683.490391] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] self.wait() [ 683.490391] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 683.490391] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] self[:] = self._gt.wait() [ 683.490391] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 683.490391] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] return self._exit_event.wait() [ 683.490391] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 683.490391] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] current.throw(*self._exc) [ 683.490391] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.490391] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] result = function(*args, **kwargs) [ 683.490669] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 683.490669] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] return func(*args, **kwargs) [ 683.490669] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 683.490669] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] raise e [ 683.490669] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.490669] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] nwinfo = self.network_api.allocate_for_instance( [ 683.490669] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 683.490669] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] created_port_ids = self._update_ports_for_instance( [ 683.490669] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 683.490669] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] with excutils.save_and_reraise_exception(): [ 683.490669] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.490669] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] self.force_reraise() [ 683.490669] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.490953] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] raise self.value [ 683.490953] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 683.490953] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] updated_port = self._update_port( [ 683.490953] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.490953] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] _ensure_no_port_binding_failure(port) [ 683.490953] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.490953] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] raise exception.PortBindingFailed(port_id=port['id']) [ 683.490953] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] nova.exception.PortBindingFailed: Binding failed for port e46c02d0-3a45-4099-8006-a529a1c1f2df, please check neutron logs for more information. [ 683.490953] env[62952]: ERROR nova.compute.manager [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] [ 683.490953] env[62952]: DEBUG nova.compute.utils [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Binding failed for port e46c02d0-3a45-4099-8006-a529a1c1f2df, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 683.495811] env[62952]: DEBUG nova.compute.manager [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Build of instance f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83 was re-scheduled: Binding failed for port e46c02d0-3a45-4099-8006-a529a1c1f2df, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 683.496350] env[62952]: DEBUG nova.compute.manager [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 683.497275] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Acquiring lock "refresh_cache-f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.497469] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Acquired lock "refresh_cache-f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.497726] env[62952]: DEBUG nova.network.neutron [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 683.499600] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.126s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.602011] env[62952]: INFO nova.compute.manager [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] [instance: 4eb9ea32-b2eb-440c-871b-6942590d9c01] Took 1.03 seconds to deallocate network for instance. [ 683.623624] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Releasing lock "refresh_cache-dc025711-45be-49b6-a765-f7687b94e6a1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.624067] env[62952]: DEBUG nova.compute.manager [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 683.624267] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 683.625213] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-208d1762-5205-4e6c-bf75-2ac385b44d4e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.632857] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 683.633139] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc9de1de-17fb-4c5a-b1b2-3cbec6e6a26c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.639770] env[62952]: DEBUG oslo_vmware.api [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Waiting for the task: (returnval){ [ 683.639770] env[62952]: value = "task-1263062" [ 683.639770] env[62952]: _type = "Task" [ 683.639770] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.650183] env[62952]: DEBUG oslo_vmware.api [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263062, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.655260] env[62952]: DEBUG nova.compute.manager [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 683.870986] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Acquiring lock "ebff0c87-8e58-4a76-a350-43a3d660debd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.871274] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Lock "ebff0c87-8e58-4a76-a350-43a3d660debd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.883508] env[62952]: DEBUG nova.compute.manager [None req-7062643b-666b-4a62-8457-6ee195941655 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Instance disappeared during snapshot {{(pid=62952) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 684.034613] env[62952]: DEBUG nova.compute.manager [None req-7062643b-666b-4a62-8457-6ee195941655 tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Found 0 images (rotation: 2) {{(pid=62952) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 684.039233] env[62952]: DEBUG nova.network.neutron [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 684.151084] env[62952]: DEBUG oslo_vmware.api [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263062, 'name': PowerOffVM_Task, 'duration_secs': 0.117443} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.151958] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 684.152355] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 684.152624] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd73ddc6-6e96-4071-a33b-745afa9d3e14 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.171288] env[62952]: DEBUG nova.network.neutron [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.180834] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 684.181066] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 684.181244] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Deleting the datastore file [datastore1] dc025711-45be-49b6-a765-f7687b94e6a1 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 684.184759] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c6b0f15-8c5d-4473-ab07-b8bd62f96bd5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.187694] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.194901] env[62952]: DEBUG oslo_vmware.api [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Waiting for the task: (returnval){ [ 684.194901] env[62952]: value = "task-1263064" [ 684.194901] env[62952]: _type = "Task" [ 684.194901] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.205121] env[62952]: DEBUG oslo_vmware.api [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263064, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.482408] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cd6c336-627d-4a08-a168-f7a1cd7429cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.491034] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2122eeee-eee8-41cd-8ed4-e4fe2d8052d6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.528183] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8385519c-363d-489e-bbb4-028b9254df60 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.536956] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1625b25-b0f5-4258-9b9e-0b612caa566b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.554734] env[62952]: DEBUG nova.compute.provider_tree [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.633446] env[62952]: INFO nova.scheduler.client.report [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Deleted allocations for instance 4eb9ea32-b2eb-440c-871b-6942590d9c01 [ 684.674838] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Releasing lock "refresh_cache-f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.675404] env[62952]: DEBUG nova.compute.manager [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 684.675404] env[62952]: DEBUG nova.compute.manager [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 684.675404] env[62952]: DEBUG nova.network.neutron [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 684.709135] env[62952]: DEBUG oslo_vmware.api [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Task: {'id': task-1263064, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.367944} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.710031] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 684.710031] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 684.710031] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 684.710031] env[62952]: INFO nova.compute.manager [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Took 1.09 seconds to destroy the instance on the hypervisor. [ 684.710296] env[62952]: DEBUG oslo.service.loopingcall [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 684.710341] env[62952]: DEBUG nova.compute.manager [-] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 684.710433] env[62952]: DEBUG nova.network.neutron [-] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 684.838571] env[62952]: DEBUG nova.network.neutron [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 684.852552] env[62952]: DEBUG nova.network.neutron [-] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.058399] env[62952]: DEBUG nova.scheduler.client.report [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 685.146959] env[62952]: DEBUG oslo_concurrency.lockutils [None req-14b717d6-8a71-487d-bd60-b715fb6ddc46 tempest-ImagesOneServerNegativeTestJSON-1882530528 tempest-ImagesOneServerNegativeTestJSON-1882530528-project-member] Lock "4eb9ea32-b2eb-440c-871b-6942590d9c01" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.143s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.341466] env[62952]: DEBUG nova.network.neutron [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.355118] env[62952]: DEBUG nova.network.neutron [-] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.562877] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.063s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.563556] env[62952]: ERROR nova.compute.manager [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2f35b2b4-e110-4bb2-a3c3-0579a983f5a7, please check neutron logs for more information. [ 685.563556] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Traceback (most recent call last): [ 685.563556] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 685.563556] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] self.driver.spawn(context, instance, image_meta, [ 685.563556] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 685.563556] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 685.563556] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 685.563556] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] vm_ref = self.build_virtual_machine(instance, [ 685.563556] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 685.563556] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] vif_infos = vmwarevif.get_vif_info(self._session, [ 685.563556] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 685.564029] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] for vif in network_info: [ 685.564029] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 685.564029] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] return self._sync_wrapper(fn, *args, **kwargs) [ 685.564029] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 685.564029] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] self.wait() [ 685.564029] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 685.564029] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] self[:] = self._gt.wait() [ 685.564029] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 685.564029] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] return self._exit_event.wait() [ 685.564029] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 685.564029] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] current.throw(*self._exc) [ 685.564029] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.564029] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] result = function(*args, **kwargs) [ 685.564563] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 685.564563] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] return func(*args, **kwargs) [ 685.564563] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 685.564563] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] raise e [ 685.564563] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.564563] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] nwinfo = self.network_api.allocate_for_instance( [ 685.564563] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 685.564563] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] created_port_ids = self._update_ports_for_instance( [ 685.564563] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 685.564563] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] with excutils.save_and_reraise_exception(): [ 685.564563] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.564563] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] self.force_reraise() [ 685.564563] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.565082] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] raise self.value [ 685.565082] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 685.565082] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] updated_port = self._update_port( [ 685.565082] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.565082] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] _ensure_no_port_binding_failure(port) [ 685.565082] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.565082] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] raise exception.PortBindingFailed(port_id=port['id']) [ 685.565082] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] nova.exception.PortBindingFailed: Binding failed for port 2f35b2b4-e110-4bb2-a3c3-0579a983f5a7, please check neutron logs for more information. [ 685.565082] env[62952]: ERROR nova.compute.manager [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] [ 685.565082] env[62952]: DEBUG nova.compute.utils [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Binding failed for port 2f35b2b4-e110-4bb2-a3c3-0579a983f5a7, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 685.566439] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.695s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.570061] env[62952]: DEBUG nova.compute.manager [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Build of instance 6ac19031-ac4e-4061-bbfd-75bd5c90803a was re-scheduled: Binding failed for port 2f35b2b4-e110-4bb2-a3c3-0579a983f5a7, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 685.571927] env[62952]: DEBUG nova.compute.manager [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 685.572717] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Acquiring lock "refresh_cache-6ac19031-ac4e-4061-bbfd-75bd5c90803a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.573486] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Acquired lock "refresh_cache-6ac19031-ac4e-4061-bbfd-75bd5c90803a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.573701] env[62952]: DEBUG nova.network.neutron [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 685.649675] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 685.844750] env[62952]: INFO nova.compute.manager [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] [instance: f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83] Took 1.17 seconds to deallocate network for instance. [ 685.858332] env[62952]: INFO nova.compute.manager [-] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Took 1.15 seconds to deallocate network for instance. [ 686.104053] env[62952]: DEBUG nova.network.neutron [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.172190] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.245212] env[62952]: DEBUG nova.network.neutron [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.366545] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.635727] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae92f106-a15a-403e-94c0-7be83266833d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.645088] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ffe248-0387-494a-b48f-883a34166bd3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.677299] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c986aa-2d57-4e0d-835a-196da02701d1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.685024] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d110737-985c-4ba5-a0b4-689b52f085dd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.701544] env[62952]: DEBUG nova.compute.provider_tree [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.749684] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Releasing lock "refresh_cache-6ac19031-ac4e-4061-bbfd-75bd5c90803a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.749934] env[62952]: DEBUG nova.compute.manager [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 686.750238] env[62952]: DEBUG nova.compute.manager [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 686.750336] env[62952]: DEBUG nova.network.neutron [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 686.782989] env[62952]: DEBUG nova.network.neutron [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.900916] env[62952]: INFO nova.scheduler.client.report [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Deleted allocations for instance f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83 [ 687.206546] env[62952]: DEBUG nova.scheduler.client.report [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 687.286210] env[62952]: DEBUG nova.network.neutron [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.412242] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e7e172-805f-4413-940c-de9e356502de tempest-FloatingIPsAssociationNegativeTestJSON-1659631625 tempest-FloatingIPsAssociationNegativeTestJSON-1659631625-project-member] Lock "f8e0cee3-11f9-4e1c-9869-9d4ee14c7d83" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.326s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.718263] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.152s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.719065] env[62952]: ERROR nova.compute.manager [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1d9877c9-1496-4654-a090-78b82fabb64e, please check neutron logs for more information. [ 687.719065] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Traceback (most recent call last): [ 687.719065] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 687.719065] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] self.driver.spawn(context, instance, image_meta, [ 687.719065] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 687.719065] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.719065] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.719065] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] vm_ref = self.build_virtual_machine(instance, [ 687.719065] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.719065] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.719065] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.720847] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] for vif in network_info: [ 687.720847] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.720847] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] return self._sync_wrapper(fn, *args, **kwargs) [ 687.720847] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.720847] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] self.wait() [ 687.720847] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.720847] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] self[:] = self._gt.wait() [ 687.720847] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.720847] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] return self._exit_event.wait() [ 687.720847] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 687.720847] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] current.throw(*self._exc) [ 687.720847] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.720847] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] result = function(*args, **kwargs) [ 687.721167] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.721167] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] return func(*args, **kwargs) [ 687.721167] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.721167] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] raise e [ 687.721167] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.721167] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] nwinfo = self.network_api.allocate_for_instance( [ 687.721167] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 687.721167] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] created_port_ids = self._update_ports_for_instance( [ 687.721167] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 687.721167] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] with excutils.save_and_reraise_exception(): [ 687.721167] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.721167] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] self.force_reraise() [ 687.721167] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.721905] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] raise self.value [ 687.721905] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 687.721905] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] updated_port = self._update_port( [ 687.721905] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.721905] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] _ensure_no_port_binding_failure(port) [ 687.721905] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.721905] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] raise exception.PortBindingFailed(port_id=port['id']) [ 687.721905] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] nova.exception.PortBindingFailed: Binding failed for port 1d9877c9-1496-4654-a090-78b82fabb64e, please check neutron logs for more information. [ 687.721905] env[62952]: ERROR nova.compute.manager [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] [ 687.721905] env[62952]: DEBUG nova.compute.utils [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Binding failed for port 1d9877c9-1496-4654-a090-78b82fabb64e, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 687.722220] env[62952]: DEBUG oslo_concurrency.lockutils [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.978s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.723672] env[62952]: INFO nova.compute.claims [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.726731] env[62952]: DEBUG nova.compute.manager [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Build of instance 0fd60040-c3ab-491f-a53e-f3297657367e was re-scheduled: Binding failed for port 1d9877c9-1496-4654-a090-78b82fabb64e, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 687.727214] env[62952]: DEBUG nova.compute.manager [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 687.727451] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Acquiring lock "refresh_cache-0fd60040-c3ab-491f-a53e-f3297657367e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.727602] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Acquired lock "refresh_cache-0fd60040-c3ab-491f-a53e-f3297657367e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.727796] env[62952]: DEBUG nova.network.neutron [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 687.788984] env[62952]: INFO nova.compute.manager [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] [instance: 6ac19031-ac4e-4061-bbfd-75bd5c90803a] Took 1.04 seconds to deallocate network for instance. [ 687.916723] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 688.201276] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Acquiring lock "1ed9de84-6178-4173-84ad-dc115be7a1fa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.201555] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Lock "1ed9de84-6178-4173-84ad-dc115be7a1fa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.253884] env[62952]: DEBUG nova.network.neutron [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 688.445529] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.521293] env[62952]: DEBUG nova.network.neutron [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.831379] env[62952]: INFO nova.scheduler.client.report [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Deleted allocations for instance 6ac19031-ac4e-4061-bbfd-75bd5c90803a [ 689.025982] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Releasing lock "refresh_cache-0fd60040-c3ab-491f-a53e-f3297657367e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.025982] env[62952]: DEBUG nova.compute.manager [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 689.025982] env[62952]: DEBUG nova.compute.manager [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 689.025982] env[62952]: DEBUG nova.network.neutron [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 689.044395] env[62952]: DEBUG nova.network.neutron [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 689.205025] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a22ef02-97ff-48ff-85b6-fe3d6f8308f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.215769] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba62016b-011a-49f6-a2fe-001d030c0464 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.257075] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f44c18cb-b76d-48af-a88d-449b1f0e51b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.264318] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e6d449-b8f2-4d6d-8871-1ce803d6e38f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.277969] env[62952]: DEBUG nova.compute.provider_tree [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.343550] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19f11c7e-85bd-451b-a8d2-4381c65416dd tempest-ServersTestJSON-552840931 tempest-ServersTestJSON-552840931-project-member] Lock "6ac19031-ac4e-4061-bbfd-75bd5c90803a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.115s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.549054] env[62952]: DEBUG nova.network.neutron [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.783099] env[62952]: DEBUG nova.scheduler.client.report [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 689.847173] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 690.054076] env[62952]: INFO nova.compute.manager [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 0fd60040-c3ab-491f-a53e-f3297657367e] Took 1.03 seconds to deallocate network for instance. [ 690.293197] env[62952]: DEBUG oslo_concurrency.lockutils [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.293197] env[62952]: DEBUG nova.compute.manager [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 690.294925] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.485s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.296601] env[62952]: INFO nova.compute.claims [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.377383] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.800739] env[62952]: DEBUG nova.compute.utils [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 690.807571] env[62952]: DEBUG nova.compute.manager [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Not allocating networking since 'none' was specified. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 691.091931] env[62952]: INFO nova.scheduler.client.report [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Deleted allocations for instance 0fd60040-c3ab-491f-a53e-f3297657367e [ 691.306350] env[62952]: DEBUG nova.compute.manager [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 691.602829] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2771e146-4a70-4578-8703-d86b4325c6c2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Lock "0fd60040-c3ab-491f-a53e-f3297657367e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.292s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.759504] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-238d534e-e540-4b08-a679-08ffce203d1b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.768667] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3b33b8-9d8c-4d4a-8aaf-76621e27f3ae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.805607] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8200be62-4f1b-4368-b950-6671ba05eb8a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.816929] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31908a0b-6bfc-4fa3-b840-fe6f9a4b4ab0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.834099] env[62952]: DEBUG nova.compute.provider_tree [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.106847] env[62952]: DEBUG nova.compute.manager [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 692.323345] env[62952]: DEBUG nova.compute.manager [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 692.340590] env[62952]: DEBUG nova.scheduler.client.report [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.354980] env[62952]: DEBUG nova.virt.hardware [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 692.356197] env[62952]: DEBUG nova.virt.hardware [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 692.356197] env[62952]: DEBUG nova.virt.hardware [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 692.356197] env[62952]: DEBUG nova.virt.hardware [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 692.356197] env[62952]: DEBUG nova.virt.hardware [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 692.356197] env[62952]: DEBUG nova.virt.hardware [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 692.356197] env[62952]: DEBUG nova.virt.hardware [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 692.356431] env[62952]: DEBUG nova.virt.hardware [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 692.356431] env[62952]: DEBUG nova.virt.hardware [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 692.356492] env[62952]: DEBUG nova.virt.hardware [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 692.356745] env[62952]: DEBUG nova.virt.hardware [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 692.357548] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1027c712-b81e-4054-9b40-c92d6c1a096b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.367732] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32557c31-3a59-40e7-9285-82b75e8418b4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.385623] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 692.395017] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Creating folder: Project (de660f7b35c1444a98d4ee9181229386). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 692.395017] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-59ada5fb-ce33-4a0b-8c37-e602a6f0d785 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.407535] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Created folder: Project (de660f7b35c1444a98d4ee9181229386) in parent group-v271811. [ 692.407739] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Creating folder: Instances. Parent ref: group-v271819. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 692.408712] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d8b7c471-5e4d-490a-828b-31795cb403f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.417181] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Created folder: Instances in parent group-v271819. [ 692.417399] env[62952]: DEBUG oslo.service.loopingcall [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 692.422570] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 692.422570] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d55d73aa-cc9d-4557-b3c7-e8a9a2183e7b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.437310] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 692.437310] env[62952]: value = "task-1263067" [ 692.437310] env[62952]: _type = "Task" [ 692.437310] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.445258] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263067, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.633345] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.847580] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.552s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.847580] env[62952]: DEBUG nova.compute.manager [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 692.850460] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.539s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.949553] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263067, 'name': CreateVM_Task, 'duration_secs': 0.286996} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.949744] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 692.950643] env[62952]: DEBUG oslo_vmware.service [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d717cc45-c95e-458f-84e0-67c472cf5448 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.956937] env[62952]: DEBUG oslo_concurrency.lockutils [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.957125] env[62952]: DEBUG oslo_concurrency.lockutils [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.957488] env[62952]: DEBUG oslo_concurrency.lockutils [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 692.957738] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c822bde2-f912-42f2-8582-4068a86d5afe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.963986] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 692.963986] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52398328-3d93-05b6-8264-562d466ae25d" [ 692.963986] env[62952]: _type = "Task" [ 692.963986] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.970132] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52398328-3d93-05b6-8264-562d466ae25d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.357469] env[62952]: DEBUG nova.compute.utils [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 693.366612] env[62952]: DEBUG nova.compute.manager [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 693.366612] env[62952]: DEBUG nova.network.neutron [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 693.448837] env[62952]: DEBUG nova.policy [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3275413eddd9420da76c7a573199b45e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '134d023439644c59a18978a8e2bd9155', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 693.480407] env[62952]: DEBUG oslo_concurrency.lockutils [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.480571] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 693.480915] env[62952]: DEBUG oslo_concurrency.lockutils [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.481089] env[62952]: DEBUG oslo_concurrency.lockutils [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.481285] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 693.481624] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ba636895-f887-4165-aacb-094722cc0a2e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.499106] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 693.500401] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 693.500401] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa109d15-7b0a-4fdd-9656-5520d4439bf2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.509860] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-942d949e-f2c5-4d7e-b1fc-63e1ecf305ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.515137] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 693.515137] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ce26bf-3c60-a1b4-aa70-7e704d373ce4" [ 693.515137] env[62952]: _type = "Task" [ 693.515137] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.522933] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ce26bf-3c60-a1b4-aa70-7e704d373ce4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.783103] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35546e1e-6e87-423c-b6aa-e385144e4457 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.790325] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef30485f-6a8a-48a9-adab-3fbdda152866 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.821177] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a89334-b441-46c5-ae0a-4a637e902491 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.829080] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a377ff02-b8d4-483f-b3e2-96403c2ae538 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.844424] env[62952]: DEBUG nova.compute.provider_tree [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.866499] env[62952]: DEBUG nova.compute.manager [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 693.912453] env[62952]: DEBUG nova.network.neutron [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Successfully created port: 2e9ca01d-0d4c-484c-a227-21069ad4c486 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 694.024869] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Preparing fetch location {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 694.025148] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Creating directory with path [datastore2] vmware_temp/53961a6c-0c2d-41bd-91f0-a186541c532c/119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 694.025384] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9dac9edb-6533-4984-83c9-27d46985f6e8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.048186] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Created directory with path [datastore2] vmware_temp/53961a6c-0c2d-41bd-91f0-a186541c532c/119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 694.048406] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Fetch image to [datastore2] vmware_temp/53961a6c-0c2d-41bd-91f0-a186541c532c/119dd2b4-b6de-42d5-adba-db818e4c2cd7/tmp-sparse.vmdk {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 694.048579] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Downloading image file data 119dd2b4-b6de-42d5-adba-db818e4c2cd7 to [datastore2] vmware_temp/53961a6c-0c2d-41bd-91f0-a186541c532c/119dd2b4-b6de-42d5-adba-db818e4c2cd7/tmp-sparse.vmdk on the data store datastore2 {{(pid=62952) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 694.049401] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-318d459f-43fb-405a-9adc-13e1649643da {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.059167] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc296a6-45da-48c8-b341-9f20337509c0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.068211] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-470e97a1-c6dc-4c4b-ade4-904db8c8d932 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.102924] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3092232-f114-4c20-80ba-915d0e2c2de5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.109267] env[62952]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-32ad9a15-8b59-4ae7-9d62-bd255ef87435 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.144067] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Downloading image file data 119dd2b4-b6de-42d5-adba-db818e4c2cd7 to the data store datastore2 {{(pid=62952) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 694.223506] env[62952]: DEBUG oslo_vmware.rw_handles [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/53961a6c-0c2d-41bd-91f0-a186541c532c/119dd2b4-b6de-42d5-adba-db818e4c2cd7/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62952) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 694.348969] env[62952]: DEBUG nova.scheduler.client.report [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.854882] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.004s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.855944] env[62952]: ERROR nova.compute.manager [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c97da0ad-3a77-4848-ac45-978faf4e8ce6, please check neutron logs for more information. [ 694.855944] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Traceback (most recent call last): [ 694.855944] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 694.855944] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] self.driver.spawn(context, instance, image_meta, [ 694.855944] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 694.855944] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.855944] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.855944] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] vm_ref = self.build_virtual_machine(instance, [ 694.855944] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.855944] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.855944] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.856333] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] for vif in network_info: [ 694.856333] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.856333] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] return self._sync_wrapper(fn, *args, **kwargs) [ 694.856333] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.856333] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] self.wait() [ 694.856333] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.856333] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] self[:] = self._gt.wait() [ 694.856333] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.856333] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] return self._exit_event.wait() [ 694.856333] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 694.856333] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] current.throw(*self._exc) [ 694.856333] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.856333] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] result = function(*args, **kwargs) [ 694.856685] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.856685] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] return func(*args, **kwargs) [ 694.856685] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.856685] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] raise e [ 694.856685] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.856685] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] nwinfo = self.network_api.allocate_for_instance( [ 694.856685] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.856685] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] created_port_ids = self._update_ports_for_instance( [ 694.856685] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.856685] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] with excutils.save_and_reraise_exception(): [ 694.856685] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.856685] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] self.force_reraise() [ 694.856685] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.857038] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] raise self.value [ 694.857038] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.857038] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] updated_port = self._update_port( [ 694.857038] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.857038] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] _ensure_no_port_binding_failure(port) [ 694.857038] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.857038] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] raise exception.PortBindingFailed(port_id=port['id']) [ 694.857038] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] nova.exception.PortBindingFailed: Binding failed for port c97da0ad-3a77-4848-ac45-978faf4e8ce6, please check neutron logs for more information. [ 694.857038] env[62952]: ERROR nova.compute.manager [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] [ 694.857038] env[62952]: DEBUG nova.compute.utils [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Binding failed for port c97da0ad-3a77-4848-ac45-978faf4e8ce6, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 694.858150] env[62952]: DEBUG oslo_concurrency.lockutils [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.633s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.860460] env[62952]: INFO nova.compute.claims [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 694.866541] env[62952]: DEBUG nova.compute.manager [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Build of instance 4c6f81ba-85f5-4ad8-a3c8-57867c613baa was re-scheduled: Binding failed for port c97da0ad-3a77-4848-ac45-978faf4e8ce6, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 694.866541] env[62952]: DEBUG nova.compute.manager [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 694.866541] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Acquiring lock "refresh_cache-4c6f81ba-85f5-4ad8-a3c8-57867c613baa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.866541] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Acquired lock "refresh_cache-4c6f81ba-85f5-4ad8-a3c8-57867c613baa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.866883] env[62952]: DEBUG nova.network.neutron [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 694.880026] env[62952]: DEBUG nova.compute.manager [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 694.886711] env[62952]: DEBUG oslo_vmware.rw_handles [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Completed reading data from the image iterator. {{(pid=62952) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 694.887032] env[62952]: DEBUG oslo_vmware.rw_handles [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/53961a6c-0c2d-41bd-91f0-a186541c532c/119dd2b4-b6de-42d5-adba-db818e4c2cd7/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 694.913530] env[62952]: DEBUG nova.virt.hardware [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 694.913795] env[62952]: DEBUG nova.virt.hardware [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 694.914037] env[62952]: DEBUG nova.virt.hardware [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 694.914290] env[62952]: DEBUG nova.virt.hardware [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 694.914446] env[62952]: DEBUG nova.virt.hardware [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 694.914588] env[62952]: DEBUG nova.virt.hardware [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 694.914834] env[62952]: DEBUG nova.virt.hardware [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 694.914996] env[62952]: DEBUG nova.virt.hardware [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 694.915311] env[62952]: DEBUG nova.virt.hardware [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 694.915392] env[62952]: DEBUG nova.virt.hardware [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 694.915494] env[62952]: DEBUG nova.virt.hardware [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 694.916380] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caa982b9-ae39-433f-8326-05bd03013d6b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.925506] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54f1dba7-eb7f-4ffd-8bfd-4caf1291ab1a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.032493] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Downloaded image file data 119dd2b4-b6de-42d5-adba-db818e4c2cd7 to vmware_temp/53961a6c-0c2d-41bd-91f0-a186541c532c/119dd2b4-b6de-42d5-adba-db818e4c2cd7/tmp-sparse.vmdk on the data store datastore2 {{(pid=62952) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 695.035040] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Caching image {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 695.035040] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Copying Virtual Disk [datastore2] vmware_temp/53961a6c-0c2d-41bd-91f0-a186541c532c/119dd2b4-b6de-42d5-adba-db818e4c2cd7/tmp-sparse.vmdk to [datastore2] vmware_temp/53961a6c-0c2d-41bd-91f0-a186541c532c/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 695.036552] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5abee551-8cc2-44b5-81b4-6bcfe52876af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.045851] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 695.045851] env[62952]: value = "task-1263068" [ 695.045851] env[62952]: _type = "Task" [ 695.045851] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.057326] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263068, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.326237] env[62952]: DEBUG nova.compute.manager [req-53c9316f-8883-45ce-978e-01e995825918 req-8dd4b1c5-2e8f-49f8-998c-60e9f1369892 service nova] [instance: d3559857-e983-4790-8283-05aa02437723] Received event network-changed-2e9ca01d-0d4c-484c-a227-21069ad4c486 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 695.326237] env[62952]: DEBUG nova.compute.manager [req-53c9316f-8883-45ce-978e-01e995825918 req-8dd4b1c5-2e8f-49f8-998c-60e9f1369892 service nova] [instance: d3559857-e983-4790-8283-05aa02437723] Refreshing instance network info cache due to event network-changed-2e9ca01d-0d4c-484c-a227-21069ad4c486. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 695.326237] env[62952]: DEBUG oslo_concurrency.lockutils [req-53c9316f-8883-45ce-978e-01e995825918 req-8dd4b1c5-2e8f-49f8-998c-60e9f1369892 service nova] Acquiring lock "refresh_cache-d3559857-e983-4790-8283-05aa02437723" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.326237] env[62952]: DEBUG oslo_concurrency.lockutils [req-53c9316f-8883-45ce-978e-01e995825918 req-8dd4b1c5-2e8f-49f8-998c-60e9f1369892 service nova] Acquired lock "refresh_cache-d3559857-e983-4790-8283-05aa02437723" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.327030] env[62952]: DEBUG nova.network.neutron [req-53c9316f-8883-45ce-978e-01e995825918 req-8dd4b1c5-2e8f-49f8-998c-60e9f1369892 service nova] [instance: d3559857-e983-4790-8283-05aa02437723] Refreshing network info cache for port 2e9ca01d-0d4c-484c-a227-21069ad4c486 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 695.402382] env[62952]: DEBUG nova.network.neutron [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.559532] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263068, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.586586] env[62952]: DEBUG nova.network.neutron [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.629776] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "1d6447ab-dc07-459f-9124-d881e0da7cf1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.629854] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "1d6447ab-dc07-459f-9124-d881e0da7cf1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.678166] env[62952]: ERROR nova.compute.manager [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2e9ca01d-0d4c-484c-a227-21069ad4c486, please check neutron logs for more information. [ 695.678166] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 695.678166] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.678166] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 695.678166] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 695.678166] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 695.678166] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 695.678166] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 695.678166] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.678166] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 695.678166] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.678166] env[62952]: ERROR nova.compute.manager raise self.value [ 695.678166] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 695.678166] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 695.678166] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.678166] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 695.678887] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.678887] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 695.678887] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2e9ca01d-0d4c-484c-a227-21069ad4c486, please check neutron logs for more information. [ 695.678887] env[62952]: ERROR nova.compute.manager [ 695.678887] env[62952]: Traceback (most recent call last): [ 695.678887] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 695.678887] env[62952]: listener.cb(fileno) [ 695.678887] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.678887] env[62952]: result = function(*args, **kwargs) [ 695.678887] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.678887] env[62952]: return func(*args, **kwargs) [ 695.678887] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.678887] env[62952]: raise e [ 695.678887] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.678887] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 695.678887] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 695.678887] env[62952]: created_port_ids = self._update_ports_for_instance( [ 695.678887] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 695.678887] env[62952]: with excutils.save_and_reraise_exception(): [ 695.678887] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.678887] env[62952]: self.force_reraise() [ 695.678887] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.678887] env[62952]: raise self.value [ 695.678887] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 695.678887] env[62952]: updated_port = self._update_port( [ 695.678887] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.678887] env[62952]: _ensure_no_port_binding_failure(port) [ 695.678887] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.678887] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 695.680428] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 2e9ca01d-0d4c-484c-a227-21069ad4c486, please check neutron logs for more information. [ 695.680428] env[62952]: Removing descriptor: 19 [ 695.680428] env[62952]: ERROR nova.compute.manager [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2e9ca01d-0d4c-484c-a227-21069ad4c486, please check neutron logs for more information. [ 695.680428] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] Traceback (most recent call last): [ 695.680428] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 695.680428] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] yield resources [ 695.680428] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 695.680428] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] self.driver.spawn(context, instance, image_meta, [ 695.680428] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 695.680428] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] self._vmops.spawn(context, instance, image_meta, injected_files, [ 695.680428] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 695.680428] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] vm_ref = self.build_virtual_machine(instance, [ 695.681510] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 695.681510] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] vif_infos = vmwarevif.get_vif_info(self._session, [ 695.681510] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 695.681510] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] for vif in network_info: [ 695.681510] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 695.681510] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] return self._sync_wrapper(fn, *args, **kwargs) [ 695.681510] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 695.681510] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] self.wait() [ 695.681510] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 695.681510] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] self[:] = self._gt.wait() [ 695.681510] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 695.681510] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] return self._exit_event.wait() [ 695.681510] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 695.682447] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] result = hub.switch() [ 695.682447] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 695.682447] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] return self.greenlet.switch() [ 695.682447] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.682447] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] result = function(*args, **kwargs) [ 695.682447] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.682447] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] return func(*args, **kwargs) [ 695.682447] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.682447] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] raise e [ 695.682447] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.682447] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] nwinfo = self.network_api.allocate_for_instance( [ 695.682447] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 695.682447] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] created_port_ids = self._update_ports_for_instance( [ 695.683354] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 695.683354] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] with excutils.save_and_reraise_exception(): [ 695.683354] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.683354] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] self.force_reraise() [ 695.683354] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.683354] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] raise self.value [ 695.683354] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 695.683354] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] updated_port = self._update_port( [ 695.683354] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.683354] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] _ensure_no_port_binding_failure(port) [ 695.683354] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.683354] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] raise exception.PortBindingFailed(port_id=port['id']) [ 695.684009] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] nova.exception.PortBindingFailed: Binding failed for port 2e9ca01d-0d4c-484c-a227-21069ad4c486, please check neutron logs for more information. [ 695.684009] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] [ 695.684009] env[62952]: INFO nova.compute.manager [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Terminating instance [ 695.688774] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Acquiring lock "refresh_cache-d3559857-e983-4790-8283-05aa02437723" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.689288] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "9c5167eb-3a15-47ff-923e-c83cc3529cc8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.689547] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "9c5167eb-3a15-47ff-923e-c83cc3529cc8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.849720] env[62952]: DEBUG nova.network.neutron [req-53c9316f-8883-45ce-978e-01e995825918 req-8dd4b1c5-2e8f-49f8-998c-60e9f1369892 service nova] [instance: d3559857-e983-4790-8283-05aa02437723] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.968193] env[62952]: DEBUG nova.network.neutron [req-53c9316f-8883-45ce-978e-01e995825918 req-8dd4b1c5-2e8f-49f8-998c-60e9f1369892 service nova] [instance: d3559857-e983-4790-8283-05aa02437723] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.057387] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263068, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.688222} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.060013] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Copied Virtual Disk [datastore2] vmware_temp/53961a6c-0c2d-41bd-91f0-a186541c532c/119dd2b4-b6de-42d5-adba-db818e4c2cd7/tmp-sparse.vmdk to [datastore2] vmware_temp/53961a6c-0c2d-41bd-91f0-a186541c532c/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 696.060374] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Deleting the datastore file [datastore2] vmware_temp/53961a6c-0c2d-41bd-91f0-a186541c532c/119dd2b4-b6de-42d5-adba-db818e4c2cd7/tmp-sparse.vmdk {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 696.060917] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2fa85e3d-13ce-41cf-b8a9-d526878f1a10 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.068201] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 696.068201] env[62952]: value = "task-1263069" [ 696.068201] env[62952]: _type = "Task" [ 696.068201] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.077105] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263069, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.090036] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Releasing lock "refresh_cache-4c6f81ba-85f5-4ad8-a3c8-57867c613baa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.090036] env[62952]: DEBUG nova.compute.manager [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 696.090036] env[62952]: DEBUG nova.compute.manager [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.090256] env[62952]: DEBUG nova.network.neutron [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 696.118385] env[62952]: DEBUG nova.network.neutron [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.356102] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904324ea-73c0-411c-a810-1701b47f0e86 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.364177] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a69a0c-7358-45f5-b06b-af4314096867 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.403247] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb85e397-fd6f-4ac2-9da8-1f9020f868e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.411738] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18b5769e-107d-4ffe-bc58-eec3ca04576a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.425327] env[62952]: DEBUG nova.compute.provider_tree [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.476110] env[62952]: DEBUG oslo_concurrency.lockutils [req-53c9316f-8883-45ce-978e-01e995825918 req-8dd4b1c5-2e8f-49f8-998c-60e9f1369892 service nova] Releasing lock "refresh_cache-d3559857-e983-4790-8283-05aa02437723" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.476110] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Acquired lock "refresh_cache-d3559857-e983-4790-8283-05aa02437723" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.476110] env[62952]: DEBUG nova.network.neutron [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 696.582311] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263069, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.030645} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.582311] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 696.582311] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Moving file from [datastore2] vmware_temp/53961a6c-0c2d-41bd-91f0-a186541c532c/119dd2b4-b6de-42d5-adba-db818e4c2cd7 to [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7. {{(pid=62952) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 696.582457] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-9ae66c3a-e996-439f-9246-dfe537cd091c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.591294] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 696.591294] env[62952]: value = "task-1263070" [ 696.591294] env[62952]: _type = "Task" [ 696.591294] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.600961] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263070, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.621290] env[62952]: DEBUG nova.network.neutron [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.928522] env[62952]: DEBUG nova.scheduler.client.report [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 696.992792] env[62952]: DEBUG nova.network.neutron [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.071702] env[62952]: DEBUG nova.network.neutron [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.104182] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263070, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024308} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.105597] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] File moved {{(pid=62952) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 697.106492] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Cleaning up location [datastore2] vmware_temp/53961a6c-0c2d-41bd-91f0-a186541c532c {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 697.106744] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Deleting the datastore file [datastore2] vmware_temp/53961a6c-0c2d-41bd-91f0-a186541c532c {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 697.106974] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3235cd68-7b4a-46c1-82a4-2b2f72812b2f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.115633] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 697.115633] env[62952]: value = "task-1263071" [ 697.115633] env[62952]: _type = "Task" [ 697.115633] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.125283] env[62952]: INFO nova.compute.manager [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] [instance: 4c6f81ba-85f5-4ad8-a3c8-57867c613baa] Took 1.03 seconds to deallocate network for instance. [ 697.127765] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263071, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.374543] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 697.374902] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 697.433412] env[62952]: DEBUG oslo_concurrency.lockutils [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.575s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.433933] env[62952]: DEBUG nova.compute.manager [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 697.436499] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.301s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.440269] env[62952]: INFO nova.compute.claims [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 697.474553] env[62952]: DEBUG nova.compute.manager [req-87d7497d-2a34-4788-8ebc-21adaefad27a req-4b2e7a14-e73f-455a-8ac3-4a24fb364e65 service nova] [instance: d3559857-e983-4790-8283-05aa02437723] Received event network-vif-deleted-2e9ca01d-0d4c-484c-a227-21069ad4c486 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 697.577953] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Releasing lock "refresh_cache-d3559857-e983-4790-8283-05aa02437723" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.577953] env[62952]: DEBUG nova.compute.manager [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 697.577953] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 697.577953] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-049943dc-6c28-4245-a47c-3af79cc5611e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.584654] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5911ae5-f40a-45ed-842f-b9cfb0e193d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.610832] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d3559857-e983-4790-8283-05aa02437723 could not be found. [ 697.611069] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 697.611208] env[62952]: INFO nova.compute.manager [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Took 0.04 seconds to destroy the instance on the hypervisor. [ 697.611452] env[62952]: DEBUG oslo.service.loopingcall [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 697.611682] env[62952]: DEBUG nova.compute.manager [-] [instance: d3559857-e983-4790-8283-05aa02437723] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 697.611771] env[62952]: DEBUG nova.network.neutron [-] [instance: d3559857-e983-4790-8283-05aa02437723] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 697.625915] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263071, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023717} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.625915] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 697.626113] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d6579b3-fa60-4cc2-bef4-38bad3689a7e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.628783] env[62952]: DEBUG nova.network.neutron [-] [instance: d3559857-e983-4790-8283-05aa02437723] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.636299] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 697.636299] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b334b6-7025-4698-8b09-6f2cf99b64b8" [ 697.636299] env[62952]: _type = "Task" [ 697.636299] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.645504] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b334b6-7025-4698-8b09-6f2cf99b64b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.882533] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 697.882533] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 697.882533] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 697.945196] env[62952]: DEBUG nova.compute.utils [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 697.948993] env[62952]: DEBUG nova.compute.manager [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Not allocating networking since 'none' was specified. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 698.133482] env[62952]: DEBUG nova.network.neutron [-] [instance: d3559857-e983-4790-8283-05aa02437723] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.147044] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b334b6-7025-4698-8b09-6f2cf99b64b8, 'name': SearchDatastore_Task, 'duration_secs': 0.008503} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.147275] env[62952]: DEBUG oslo_concurrency.lockutils [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.147523] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] f075161d-7447-4601-a06c-514fc4dd2eb9/f075161d-7447-4601-a06c-514fc4dd2eb9.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 698.147767] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e30388b7-a1ff-40e5-bc20-9bb3c8c61682 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.154113] env[62952]: INFO nova.scheduler.client.report [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Deleted allocations for instance 4c6f81ba-85f5-4ad8-a3c8-57867c613baa [ 698.162476] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 698.162476] env[62952]: value = "task-1263072" [ 698.162476] env[62952]: _type = "Task" [ 698.162476] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.170998] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263072, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.304244] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Acquiring lock "9be8f7f5-80eb-4e38-9501-e76abd7227db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.304485] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Lock "9be8f7f5-80eb-4e38-9501-e76abd7227db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.386208] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 698.386400] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: d3559857-e983-4790-8283-05aa02437723] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 698.386531] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 698.386657] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 698.423368] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "refresh_cache-dc025711-45be-49b6-a765-f7687b94e6a1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.423527] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquired lock "refresh_cache-dc025711-45be-49b6-a765-f7687b94e6a1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.423678] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Forcefully refreshing network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 698.423835] env[62952]: DEBUG nova.objects.instance [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lazy-loading 'info_cache' on Instance uuid dc025711-45be-49b6-a765-f7687b94e6a1 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 698.449778] env[62952]: DEBUG nova.compute.manager [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 698.640733] env[62952]: INFO nova.compute.manager [-] [instance: d3559857-e983-4790-8283-05aa02437723] Took 1.03 seconds to deallocate network for instance. [ 698.643324] env[62952]: DEBUG nova.compute.claims [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 698.643429] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.662850] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3e198cbe-612e-4fcd-8e81-00594b252cc2 tempest-ServersAdminTestJSON-1179080309 tempest-ServersAdminTestJSON-1179080309-project-member] Lock "4c6f81ba-85f5-4ad8-a3c8-57867c613baa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.002s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.673317] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263072, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454456} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.676029] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] f075161d-7447-4601-a06c-514fc4dd2eb9/f075161d-7447-4601-a06c-514fc4dd2eb9.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 698.676268] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 698.676755] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2e71dd08-e523-463c-b867-70901c9891ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.683264] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 698.683264] env[62952]: value = "task-1263073" [ 698.683264] env[62952]: _type = "Task" [ 698.683264] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.694014] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263073, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.846388] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3358b370-3f38-4128-b4ef-b64055a5c8b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.853806] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4a040f-4b4a-4fdd-b18d-5fed1cd333d4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.885095] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac8ba38-c5f6-46e2-a3e9-700477df7351 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.892613] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cfa129c-c130-4b56-ad7a-7c375314744f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.517268] env[62952]: DEBUG nova.compute.manager [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 699.534060] env[62952]: DEBUG nova.compute.provider_tree [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 699.539811] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263073, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063983} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.540098] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 699.540904] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a01d4c9-d67f-45e7-b7c3-4a325c218a35 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.563019] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Reconfiguring VM instance instance-00000015 to attach disk [datastore2] f075161d-7447-4601-a06c-514fc4dd2eb9/f075161d-7447-4601-a06c-514fc4dd2eb9.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 699.563019] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b7bff9a9-f49f-4687-8fcb-5d52c8b57348 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.589102] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 699.589102] env[62952]: value = "task-1263074" [ 699.589102] env[62952]: _type = "Task" [ 699.589102] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.597151] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263074, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.017896] env[62952]: DEBUG nova.compute.manager [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 700.037473] env[62952]: DEBUG nova.scheduler.client.report [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 700.052175] env[62952]: DEBUG nova.virt.hardware [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 700.052459] env[62952]: DEBUG nova.virt.hardware [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 700.052647] env[62952]: DEBUG nova.virt.hardware [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 700.052882] env[62952]: DEBUG nova.virt.hardware [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 700.054276] env[62952]: DEBUG nova.virt.hardware [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 700.054276] env[62952]: DEBUG nova.virt.hardware [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 700.055987] env[62952]: DEBUG nova.virt.hardware [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 700.055987] env[62952]: DEBUG nova.virt.hardware [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 700.055987] env[62952]: DEBUG nova.virt.hardware [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 700.056172] env[62952]: DEBUG nova.virt.hardware [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 700.056354] env[62952]: DEBUG nova.virt.hardware [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 700.057241] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c94ce8f-ee52-4a14-a298-cae14c82ffd4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.061869] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.062552] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 700.070317] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d662726-7280-4d8a-8764-7d0f2be3e938 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.085532] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 700.091192] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Creating folder: Project (f87b115a82ca4c99bda22db6a7fbe2bb). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 700.091721] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-981a4b7b-6702-44c4-9ad9-7694f53b80d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.101586] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263074, 'name': ReconfigVM_Task, 'duration_secs': 0.282609} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.101586] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Reconfigured VM instance instance-00000015 to attach disk [datastore2] f075161d-7447-4601-a06c-514fc4dd2eb9/f075161d-7447-4601-a06c-514fc4dd2eb9.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 700.102191] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b2d1e66d-33a4-4cdb-b241-5ce9ca029f68 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.105023] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Created folder: Project (f87b115a82ca4c99bda22db6a7fbe2bb) in parent group-v271811. [ 700.105023] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Creating folder: Instances. Parent ref: group-v271822. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 700.105513] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bbca2067-f761-4217-8ae9-c5573808bf23 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.109659] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 700.109659] env[62952]: value = "task-1263076" [ 700.109659] env[62952]: _type = "Task" [ 700.109659] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.120827] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Created folder: Instances in parent group-v271822. [ 700.121079] env[62952]: DEBUG oslo.service.loopingcall [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.121553] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263076, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.121761] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 700.122080] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-be9df47c-6e88-4955-af61-8a50878026e7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.141148] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 700.141148] env[62952]: value = "task-1263078" [ 700.141148] env[62952]: _type = "Task" [ 700.141148] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.150765] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263078, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.544230] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.108s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.544753] env[62952]: DEBUG nova.compute.manager [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 700.549729] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.362s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.551356] env[62952]: INFO nova.compute.claims [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 700.620969] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263076, 'name': Rename_Task, 'duration_secs': 0.138264} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.620969] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 700.620969] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2ebca2de-9236-420e-a2c5-64b85b520b19 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.622720] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.629299] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 700.629299] env[62952]: value = "task-1263079" [ 700.629299] env[62952]: _type = "Task" [ 700.629299] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.637784] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263079, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.650205] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263078, 'name': CreateVM_Task, 'duration_secs': 0.299519} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.650576] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 700.651134] env[62952]: DEBUG oslo_concurrency.lockutils [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.651426] env[62952]: DEBUG oslo_concurrency.lockutils [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.651877] env[62952]: DEBUG oslo_concurrency.lockutils [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 700.652242] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-571dfd76-9db3-4394-8846-50d232122f01 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.657354] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Waiting for the task: (returnval){ [ 700.657354] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]528f8894-1cd2-19c1-39ba-a314defbd848" [ 700.657354] env[62952]: _type = "Task" [ 700.657354] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.670417] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]528f8894-1cd2-19c1-39ba-a314defbd848, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.059903] env[62952]: DEBUG nova.compute.utils [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 701.062807] env[62952]: DEBUG nova.compute.manager [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 701.063152] env[62952]: DEBUG nova.network.neutron [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 701.117435] env[62952]: DEBUG nova.policy [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9f6f0133f4a4ac287d4d8d3f4992754', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1487c06883e444a589d9b688155dc00e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 701.127223] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Releasing lock "refresh_cache-dc025711-45be-49b6-a765-f7687b94e6a1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.127223] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Updated the network info_cache for instance {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 701.127223] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 701.127223] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 701.127223] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 701.127223] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 701.127442] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 701.127442] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 701.127442] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 701.127442] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 701.142653] env[62952]: DEBUG oslo_vmware.api [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263079, 'name': PowerOnVM_Task, 'duration_secs': 0.442718} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.142920] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 701.143143] env[62952]: INFO nova.compute.manager [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Took 8.82 seconds to spawn the instance on the hypervisor. [ 701.143409] env[62952]: DEBUG nova.compute.manager [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 701.146260] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb03c82-783a-41c5-a21b-498a70f8b5c0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.167509] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]528f8894-1cd2-19c1-39ba-a314defbd848, 'name': SearchDatastore_Task, 'duration_secs': 0.010602} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.167928] env[62952]: DEBUG oslo_concurrency.lockutils [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.168201] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 701.168448] env[62952]: DEBUG oslo_concurrency.lockutils [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.172063] env[62952]: DEBUG oslo_concurrency.lockutils [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.172063] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 701.172063] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6bf145f6-29b2-49a0-a9e4-3e1e16d75abb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.178812] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 701.178995] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 701.179781] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc5f23e0-083a-44e4-84fe-8daf746bc316 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.184950] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Waiting for the task: (returnval){ [ 701.184950] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523e162b-f3da-d843-3bfd-1b15d4eaed9d" [ 701.184950] env[62952]: _type = "Task" [ 701.184950] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.192889] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523e162b-f3da-d843-3bfd-1b15d4eaed9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.512758] env[62952]: DEBUG nova.network.neutron [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Successfully created port: d7878ac1-867d-4c80-9e77-e1080f980358 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 701.563538] env[62952]: DEBUG nova.compute.manager [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 701.631229] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.672029] env[62952]: INFO nova.compute.manager [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Took 25.94 seconds to build instance. [ 701.701366] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523e162b-f3da-d843-3bfd-1b15d4eaed9d, 'name': SearchDatastore_Task, 'duration_secs': 0.009112} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.702292] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2b5e85b-fa33-4cca-b706-e6ab4c5245f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.708239] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Waiting for the task: (returnval){ [ 701.708239] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ef259b-0b9d-4efb-8257-768215bcf463" [ 701.708239] env[62952]: _type = "Task" [ 701.708239] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.720167] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ef259b-0b9d-4efb-8257-768215bcf463, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.054224] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc2b4c1-4d94-437a-a280-c1d920348b8f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.060711] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05272f2a-0bf6-43b1-bed8-aac7a036f87b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.102406] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd14b3d-4612-444d-b913-5edb07f4d9d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.110330] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594e9af0-15be-4fa3-9b1c-cd0a5e18212b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.124875] env[62952]: DEBUG nova.compute.provider_tree [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.174841] env[62952]: DEBUG oslo_concurrency.lockutils [None req-11b11cc7-270a-4d51-bd64-b497fe9b2994 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Lock "f075161d-7447-4601-a06c-514fc4dd2eb9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.614s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.220552] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ef259b-0b9d-4efb-8257-768215bcf463, 'name': SearchDatastore_Task, 'duration_secs': 0.009241} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.220552] env[62952]: DEBUG oslo_concurrency.lockutils [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.221190] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 80a30ea8-e260-4b92-9099-be374333896f/80a30ea8-e260-4b92-9099-be374333896f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 702.221190] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-21837753-c1ff-49ea-ab1d-b0a4978c867e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.227360] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Waiting for the task: (returnval){ [ 702.227360] env[62952]: value = "task-1263080" [ 702.227360] env[62952]: _type = "Task" [ 702.227360] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.238237] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': task-1263080, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.609023] env[62952]: DEBUG nova.compute.manager [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 702.633034] env[62952]: DEBUG nova.scheduler.client.report [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 702.642036] env[62952]: DEBUG nova.virt.hardware [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 702.642292] env[62952]: DEBUG nova.virt.hardware [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 702.642441] env[62952]: DEBUG nova.virt.hardware [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 702.642637] env[62952]: DEBUG nova.virt.hardware [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 702.642790] env[62952]: DEBUG nova.virt.hardware [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 702.642930] env[62952]: DEBUG nova.virt.hardware [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 702.643181] env[62952]: DEBUG nova.virt.hardware [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 702.643352] env[62952]: DEBUG nova.virt.hardware [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 702.643573] env[62952]: DEBUG nova.virt.hardware [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 702.643763] env[62952]: DEBUG nova.virt.hardware [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 702.643985] env[62952]: DEBUG nova.virt.hardware [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 702.645451] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85babef0-2f0e-4cb4-ba56-80d09ebd7726 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.659076] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a1efe6-d0e1-47ca-9ee2-fd93a1b59ca3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.665078] env[62952]: INFO nova.compute.manager [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Rebuilding instance [ 702.691084] env[62952]: DEBUG nova.compute.manager [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 702.700744] env[62952]: DEBUG nova.compute.manager [req-2c618883-b1fa-4876-addb-c39318075aee req-5257d998-8489-4cab-96c6-0adb0474197a service nova] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Received event network-changed-d7878ac1-867d-4c80-9e77-e1080f980358 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 702.700947] env[62952]: DEBUG nova.compute.manager [req-2c618883-b1fa-4876-addb-c39318075aee req-5257d998-8489-4cab-96c6-0adb0474197a service nova] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Refreshing instance network info cache due to event network-changed-d7878ac1-867d-4c80-9e77-e1080f980358. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 702.701237] env[62952]: DEBUG oslo_concurrency.lockutils [req-2c618883-b1fa-4876-addb-c39318075aee req-5257d998-8489-4cab-96c6-0adb0474197a service nova] Acquiring lock "refresh_cache-3516272a-9ebd-413e-9b88-1a5c83b50f61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.701721] env[62952]: DEBUG oslo_concurrency.lockutils [req-2c618883-b1fa-4876-addb-c39318075aee req-5257d998-8489-4cab-96c6-0adb0474197a service nova] Acquired lock "refresh_cache-3516272a-9ebd-413e-9b88-1a5c83b50f61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.701986] env[62952]: DEBUG nova.network.neutron [req-2c618883-b1fa-4876-addb-c39318075aee req-5257d998-8489-4cab-96c6-0adb0474197a service nova] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Refreshing network info cache for port d7878ac1-867d-4c80-9e77-e1080f980358 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 702.739243] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': task-1263080, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.746978] env[62952]: DEBUG nova.compute.manager [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 702.748054] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c627fe-499c-4d18-a260-cc62c89613f8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.892119] env[62952]: ERROR nova.compute.manager [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d7878ac1-867d-4c80-9e77-e1080f980358, please check neutron logs for more information. [ 702.892119] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 702.892119] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.892119] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 702.892119] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 702.892119] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 702.892119] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 702.892119] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 702.892119] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.892119] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 702.892119] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.892119] env[62952]: ERROR nova.compute.manager raise self.value [ 702.892119] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 702.892119] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 702.892119] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.892119] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 702.892545] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.892545] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 702.892545] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d7878ac1-867d-4c80-9e77-e1080f980358, please check neutron logs for more information. [ 702.892545] env[62952]: ERROR nova.compute.manager [ 702.892648] env[62952]: Traceback (most recent call last): [ 702.892689] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 702.892689] env[62952]: listener.cb(fileno) [ 702.892689] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.892689] env[62952]: result = function(*args, **kwargs) [ 702.892689] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 702.892689] env[62952]: return func(*args, **kwargs) [ 702.892689] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.892689] env[62952]: raise e [ 702.892880] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.892880] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 702.892880] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 702.892880] env[62952]: created_port_ids = self._update_ports_for_instance( [ 702.892880] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 702.892880] env[62952]: with excutils.save_and_reraise_exception(): [ 702.892880] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.892880] env[62952]: self.force_reraise() [ 702.892880] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.892880] env[62952]: raise self.value [ 702.892880] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 702.892880] env[62952]: updated_port = self._update_port( [ 702.892880] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.892880] env[62952]: _ensure_no_port_binding_failure(port) [ 702.892880] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.892880] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 702.892880] env[62952]: nova.exception.PortBindingFailed: Binding failed for port d7878ac1-867d-4c80-9e77-e1080f980358, please check neutron logs for more information. [ 702.892880] env[62952]: Removing descriptor: 19 [ 702.893540] env[62952]: ERROR nova.compute.manager [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d7878ac1-867d-4c80-9e77-e1080f980358, please check neutron logs for more information. [ 702.893540] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Traceback (most recent call last): [ 702.893540] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 702.893540] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] yield resources [ 702.893540] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 702.893540] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] self.driver.spawn(context, instance, image_meta, [ 702.893540] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 702.893540] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] self._vmops.spawn(context, instance, image_meta, injected_files, [ 702.893540] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 702.893540] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] vm_ref = self.build_virtual_machine(instance, [ 702.893540] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 702.893790] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] vif_infos = vmwarevif.get_vif_info(self._session, [ 702.893790] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 702.893790] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] for vif in network_info: [ 702.893790] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 702.893790] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] return self._sync_wrapper(fn, *args, **kwargs) [ 702.893790] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 702.893790] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] self.wait() [ 702.893790] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 702.893790] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] self[:] = self._gt.wait() [ 702.893790] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 702.893790] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] return self._exit_event.wait() [ 702.893790] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 702.893790] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] result = hub.switch() [ 702.894087] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 702.894087] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] return self.greenlet.switch() [ 702.894087] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.894087] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] result = function(*args, **kwargs) [ 702.894087] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 702.894087] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] return func(*args, **kwargs) [ 702.894087] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.894087] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] raise e [ 702.894087] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.894087] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] nwinfo = self.network_api.allocate_for_instance( [ 702.894087] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 702.894087] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] created_port_ids = self._update_ports_for_instance( [ 702.894087] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 702.894378] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] with excutils.save_and_reraise_exception(): [ 702.894378] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.894378] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] self.force_reraise() [ 702.894378] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.894378] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] raise self.value [ 702.894378] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 702.894378] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] updated_port = self._update_port( [ 702.894378] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.894378] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] _ensure_no_port_binding_failure(port) [ 702.894378] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.894378] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] raise exception.PortBindingFailed(port_id=port['id']) [ 702.894378] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] nova.exception.PortBindingFailed: Binding failed for port d7878ac1-867d-4c80-9e77-e1080f980358, please check neutron logs for more information. [ 702.894378] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] [ 702.894712] env[62952]: INFO nova.compute.manager [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Terminating instance [ 702.896410] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "refresh_cache-3516272a-9ebd-413e-9b88-1a5c83b50f61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.141187] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.591s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.141742] env[62952]: DEBUG nova.compute.manager [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 703.144645] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.973s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.146288] env[62952]: INFO nova.compute.claims [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 703.217613] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.227358] env[62952]: DEBUG nova.network.neutron [req-2c618883-b1fa-4876-addb-c39318075aee req-5257d998-8489-4cab-96c6-0adb0474197a service nova] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 703.241409] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': task-1263080, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546165} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.241409] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 80a30ea8-e260-4b92-9099-be374333896f/80a30ea8-e260-4b92-9099-be374333896f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 703.241409] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 703.241409] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e0c800c6-74c9-49e5-b6ea-0ef6140f561a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.247522] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Waiting for the task: (returnval){ [ 703.247522] env[62952]: value = "task-1263081" [ 703.247522] env[62952]: _type = "Task" [ 703.247522] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.258657] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': task-1263081, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.262262] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 703.263032] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1aa8727a-dee5-437f-898b-61e68bf6bb42 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.269097] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 703.269097] env[62952]: value = "task-1263082" [ 703.269097] env[62952]: _type = "Task" [ 703.269097] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.277336] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263082, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.318166] env[62952]: DEBUG nova.network.neutron [req-2c618883-b1fa-4876-addb-c39318075aee req-5257d998-8489-4cab-96c6-0adb0474197a service nova] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.653722] env[62952]: DEBUG nova.compute.utils [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 703.656366] env[62952]: DEBUG nova.compute.manager [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 703.656366] env[62952]: DEBUG nova.network.neutron [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 703.732428] env[62952]: DEBUG nova.policy [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aa5d996a798b4430a9f6cb57acf987fb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f83de16a83f14816a0085bcb8d110a11', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 703.758039] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': task-1263081, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070266} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.758357] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 703.759216] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ca2ed6-861e-4b0e-9ac5-a7009c1ac6ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.780985] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Reconfiguring VM instance instance-00000017 to attach disk [datastore1] 80a30ea8-e260-4b92-9099-be374333896f/80a30ea8-e260-4b92-9099-be374333896f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 703.784167] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1a9e98a-f16a-4b17-be80-14be761b6474 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.803413] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263082, 'name': PowerOffVM_Task, 'duration_secs': 0.130678} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.804647] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 703.804881] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 703.805845] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Waiting for the task: (returnval){ [ 703.805845] env[62952]: value = "task-1263083" [ 703.805845] env[62952]: _type = "Task" [ 703.805845] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.806126] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f01e2d03-0304-461a-8db4-c77590da3e67 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.816799] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': task-1263083, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.818843] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 703.819179] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6060595a-4cf5-406d-965b-e5166bc5c5c3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.820960] env[62952]: DEBUG oslo_concurrency.lockutils [req-2c618883-b1fa-4876-addb-c39318075aee req-5257d998-8489-4cab-96c6-0adb0474197a service nova] Releasing lock "refresh_cache-3516272a-9ebd-413e-9b88-1a5c83b50f61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.821341] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquired lock "refresh_cache-3516272a-9ebd-413e-9b88-1a5c83b50f61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.821540] env[62952]: DEBUG nova.network.neutron [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 703.848246] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 703.848452] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 703.848605] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Deleting the datastore file [datastore2] f075161d-7447-4601-a06c-514fc4dd2eb9 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 703.848859] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a81d50d8-7d0f-4c18-9210-138e348a8d11 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.855347] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 703.855347] env[62952]: value = "task-1263085" [ 703.855347] env[62952]: _type = "Task" [ 703.855347] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.863418] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263085, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.155630] env[62952]: DEBUG nova.compute.manager [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 704.170594] env[62952]: DEBUG nova.network.neutron [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Successfully created port: 4c5b2393-f9f2-4e98-8c3e-9e834192ee0f {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 704.318776] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': task-1263083, 'name': ReconfigVM_Task, 'duration_secs': 0.346255} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.320524] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Reconfigured VM instance instance-00000017 to attach disk [datastore1] 80a30ea8-e260-4b92-9099-be374333896f/80a30ea8-e260-4b92-9099-be374333896f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 704.328878] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a8e21ed-51e7-4829-8a45-6d1e4f658d96 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.336035] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Waiting for the task: (returnval){ [ 704.336035] env[62952]: value = "task-1263086" [ 704.336035] env[62952]: _type = "Task" [ 704.336035] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.346476] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': task-1263086, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.347257] env[62952]: DEBUG nova.network.neutron [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 704.364455] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263085, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.092515} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.366154] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 704.366154] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 704.366154] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 704.470520] env[62952]: DEBUG nova.network.neutron [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.636381] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3218d075-703f-4aa9-bd8f-6ec175a77a49 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.647332] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f29445-093d-48c1-a6b1-b27bc84c5e97 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.685138] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f23e06eb-c589-4fb8-879a-84b2484f611f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.695160] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcbe16be-f337-4cf9-8706-beb6e9050d14 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.709028] env[62952]: DEBUG nova.compute.provider_tree [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.719256] env[62952]: DEBUG nova.compute.manager [req-4a93bbad-ac36-42d6-8768-b0a0e32ea95c req-6a3a25fc-f58b-4d7e-af2d-bf5c6f53b5ac service nova] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Received event network-vif-deleted-d7878ac1-867d-4c80-9e77-e1080f980358 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 704.845831] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': task-1263086, 'name': Rename_Task, 'duration_secs': 0.129655} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.846828] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 704.847149] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c6738487-c97d-4a62-ab90-41f231cea1fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.853794] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Waiting for the task: (returnval){ [ 704.853794] env[62952]: value = "task-1263087" [ 704.853794] env[62952]: _type = "Task" [ 704.853794] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.861903] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': task-1263087, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.974839] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Releasing lock "refresh_cache-3516272a-9ebd-413e-9b88-1a5c83b50f61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.975834] env[62952]: DEBUG nova.compute.manager [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 704.976122] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 704.976540] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f54b8a83-0f16-4c7a-a363-295ab4fc94c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.987380] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bed6f66-b9b4-4dd2-a7a6-84bf1f289a86 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.010756] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3516272a-9ebd-413e-9b88-1a5c83b50f61 could not be found. [ 705.010994] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 705.011198] env[62952]: INFO nova.compute.manager [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Took 0.04 seconds to destroy the instance on the hypervisor. [ 705.011588] env[62952]: DEBUG oslo.service.loopingcall [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 705.011654] env[62952]: DEBUG nova.compute.manager [-] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 705.011752] env[62952]: DEBUG nova.network.neutron [-] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 705.034997] env[62952]: DEBUG nova.network.neutron [-] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.188059] env[62952]: DEBUG nova.compute.manager [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 705.210729] env[62952]: DEBUG nova.virt.hardware [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 705.210998] env[62952]: DEBUG nova.virt.hardware [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 705.211167] env[62952]: DEBUG nova.virt.hardware [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 705.211345] env[62952]: DEBUG nova.virt.hardware [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 705.211486] env[62952]: DEBUG nova.virt.hardware [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 705.211632] env[62952]: DEBUG nova.virt.hardware [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 705.211822] env[62952]: DEBUG nova.virt.hardware [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 705.211973] env[62952]: DEBUG nova.virt.hardware [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 705.212265] env[62952]: DEBUG nova.virt.hardware [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 705.212435] env[62952]: DEBUG nova.virt.hardware [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 705.212605] env[62952]: DEBUG nova.virt.hardware [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 705.213479] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f2bd51-37b0-4417-9d4d-3135788f29d9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.216500] env[62952]: DEBUG nova.scheduler.client.report [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.224913] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b0bd74-0b51-45a3-b99c-9f678f71ed06 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.366964] env[62952]: DEBUG oslo_vmware.api [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': task-1263087, 'name': PowerOnVM_Task, 'duration_secs': 0.418314} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.367892] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 705.368164] env[62952]: INFO nova.compute.manager [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Took 5.35 seconds to spawn the instance on the hypervisor. [ 705.368505] env[62952]: DEBUG nova.compute.manager [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 705.369724] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77c8f13c-916c-406e-998f-f7d7113da4b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.407699] env[62952]: DEBUG nova.virt.hardware [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 705.407999] env[62952]: DEBUG nova.virt.hardware [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 705.410365] env[62952]: DEBUG nova.virt.hardware [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 705.410365] env[62952]: DEBUG nova.virt.hardware [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 705.410365] env[62952]: DEBUG nova.virt.hardware [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 705.410365] env[62952]: DEBUG nova.virt.hardware [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 705.410365] env[62952]: DEBUG nova.virt.hardware [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 705.410365] env[62952]: DEBUG nova.virt.hardware [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 705.411110] env[62952]: DEBUG nova.virt.hardware [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 705.411358] env[62952]: DEBUG nova.virt.hardware [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 705.411587] env[62952]: DEBUG nova.virt.hardware [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 705.412613] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb1e39f3-c1e1-4bbf-8628-2d65ab403388 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.422874] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e12f7e-bca0-46b8-a33a-f5408e277664 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.440939] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 705.448382] env[62952]: DEBUG oslo.service.loopingcall [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 705.449305] env[62952]: ERROR nova.compute.manager [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4c5b2393-f9f2-4e98-8c3e-9e834192ee0f, please check neutron logs for more information. [ 705.449305] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 705.449305] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.449305] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 705.449305] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 705.449305] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 705.449305] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 705.449305] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 705.449305] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.449305] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 705.449305] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.449305] env[62952]: ERROR nova.compute.manager raise self.value [ 705.449305] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 705.449305] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 705.449305] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.449305] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 705.449799] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.449799] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 705.449799] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4c5b2393-f9f2-4e98-8c3e-9e834192ee0f, please check neutron logs for more information. [ 705.449799] env[62952]: ERROR nova.compute.manager [ 705.449799] env[62952]: Traceback (most recent call last): [ 705.449799] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 705.449799] env[62952]: listener.cb(fileno) [ 705.449799] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.449799] env[62952]: result = function(*args, **kwargs) [ 705.449799] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 705.449799] env[62952]: return func(*args, **kwargs) [ 705.449799] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.449799] env[62952]: raise e [ 705.449799] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.449799] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 705.449799] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 705.449799] env[62952]: created_port_ids = self._update_ports_for_instance( [ 705.449799] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 705.449799] env[62952]: with excutils.save_and_reraise_exception(): [ 705.449799] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.449799] env[62952]: self.force_reraise() [ 705.449799] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.449799] env[62952]: raise self.value [ 705.449799] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 705.449799] env[62952]: updated_port = self._update_port( [ 705.449799] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.449799] env[62952]: _ensure_no_port_binding_failure(port) [ 705.449799] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.449799] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 705.451888] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 4c5b2393-f9f2-4e98-8c3e-9e834192ee0f, please check neutron logs for more information. [ 705.451888] env[62952]: Removing descriptor: 19 [ 705.451888] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 705.451888] env[62952]: ERROR nova.compute.manager [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4c5b2393-f9f2-4e98-8c3e-9e834192ee0f, please check neutron logs for more information. [ 705.451888] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Traceback (most recent call last): [ 705.451888] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 705.451888] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] yield resources [ 705.451888] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 705.451888] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] self.driver.spawn(context, instance, image_meta, [ 705.451888] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 705.451888] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.451888] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.452263] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] vm_ref = self.build_virtual_machine(instance, [ 705.452263] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.452263] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.452263] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.452263] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] for vif in network_info: [ 705.452263] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 705.452263] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] return self._sync_wrapper(fn, *args, **kwargs) [ 705.452263] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 705.452263] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] self.wait() [ 705.452263] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 705.452263] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] self[:] = self._gt.wait() [ 705.452263] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.452263] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] return self._exit_event.wait() [ 705.452578] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 705.452578] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] result = hub.switch() [ 705.452578] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 705.452578] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] return self.greenlet.switch() [ 705.452578] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.452578] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] result = function(*args, **kwargs) [ 705.452578] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 705.452578] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] return func(*args, **kwargs) [ 705.452578] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.452578] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] raise e [ 705.452578] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.452578] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] nwinfo = self.network_api.allocate_for_instance( [ 705.452578] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 705.452884] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] created_port_ids = self._update_ports_for_instance( [ 705.452884] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 705.452884] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] with excutils.save_and_reraise_exception(): [ 705.452884] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.452884] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] self.force_reraise() [ 705.452884] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.452884] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] raise self.value [ 705.452884] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 705.452884] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] updated_port = self._update_port( [ 705.452884] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.452884] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] _ensure_no_port_binding_failure(port) [ 705.452884] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.452884] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] raise exception.PortBindingFailed(port_id=port['id']) [ 705.453270] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] nova.exception.PortBindingFailed: Binding failed for port 4c5b2393-f9f2-4e98-8c3e-9e834192ee0f, please check neutron logs for more information. [ 705.453270] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] [ 705.453270] env[62952]: INFO nova.compute.manager [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Terminating instance [ 705.453270] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-821c6f5b-dffe-4dcd-8de5-630bdf4fc0f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.464329] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Acquiring lock "refresh_cache-11139a80-26b2-4bc6-af91-18cd3adbe8fe" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.464495] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Acquired lock "refresh_cache-11139a80-26b2-4bc6-af91-18cd3adbe8fe" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.464665] env[62952]: DEBUG nova.network.neutron [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 705.473319] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 705.473319] env[62952]: value = "task-1263088" [ 705.473319] env[62952]: _type = "Task" [ 705.473319] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.481497] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263088, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.538454] env[62952]: DEBUG nova.network.neutron [-] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.721778] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.577s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.722326] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 705.725519] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.359s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.725737] env[62952]: DEBUG nova.objects.instance [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Lazy-loading 'resources' on Instance uuid dc025711-45be-49b6-a765-f7687b94e6a1 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 705.891977] env[62952]: INFO nova.compute.manager [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Took 23.69 seconds to build instance. [ 705.986348] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263088, 'name': CreateVM_Task, 'duration_secs': 0.250671} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.986348] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 705.986348] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.986348] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.986348] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 705.986348] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5c5b4a5-bb54-4aac-8259-a2c07b275edb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.990697] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 705.990697] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520b9971-2339-1ed3-3531-b87f0df83243" [ 705.990697] env[62952]: _type = "Task" [ 705.990697] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.992754] env[62952]: DEBUG nova.network.neutron [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 706.004233] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520b9971-2339-1ed3-3531-b87f0df83243, 'name': SearchDatastore_Task, 'duration_secs': 0.008038} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.004633] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.004983] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 706.005317] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.005556] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.006169] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 706.007065] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bf230a19-be7f-4f79-a8dd-e5fdfd5d3f27 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.017074] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 706.017074] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 706.017074] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0dd95263-2c69-458f-b558-d84aae0aa097 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.021529] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 706.021529] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5280b589-7e20-47b7-3f3a-d5f04d95b286" [ 706.021529] env[62952]: _type = "Task" [ 706.021529] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.034940] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5280b589-7e20-47b7-3f3a-d5f04d95b286, 'name': SearchDatastore_Task, 'duration_secs': 0.009906} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.036619] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-047c4a2a-9912-4411-bfae-596453f9a8f7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.044034] env[62952]: INFO nova.compute.manager [-] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Took 1.03 seconds to deallocate network for instance. [ 706.044505] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 706.044505] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5284c1c3-d0f4-7725-ba8f-d12243502328" [ 706.044505] env[62952]: _type = "Task" [ 706.044505] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.045090] env[62952]: DEBUG nova.compute.claims [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 706.045155] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.053996] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5284c1c3-d0f4-7725-ba8f-d12243502328, 'name': SearchDatastore_Task, 'duration_secs': 0.007961} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.053996] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.054117] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] f075161d-7447-4601-a06c-514fc4dd2eb9/f075161d-7447-4601-a06c-514fc4dd2eb9.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 706.054287] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5817a3db-130a-46ea-9294-43422bb46e08 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.060047] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 706.060047] env[62952]: value = "task-1263089" [ 706.060047] env[62952]: _type = "Task" [ 706.060047] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.068064] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263089, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.092024] env[62952]: DEBUG nova.network.neutron [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.230843] env[62952]: DEBUG nova.compute.utils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 706.234717] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 706.235032] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 706.312742] env[62952]: DEBUG nova.policy [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dba2a0219a8e45a49cf91e11bcf203d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b141129b4bb24edd943a45cf4304c6c0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 706.396663] env[62952]: DEBUG oslo_concurrency.lockutils [None req-060d7df6-011c-441e-b735-ea07136dae72 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Lock "80a30ea8-e260-4b92-9099-be374333896f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.298s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.569625] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263089, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464965} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.572069] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] f075161d-7447-4601-a06c-514fc4dd2eb9/f075161d-7447-4601-a06c-514fc4dd2eb9.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 706.572295] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 706.572718] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-25a78ae2-69f7-4b39-a562-c066d5825862 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.579182] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 706.579182] env[62952]: value = "task-1263090" [ 706.579182] env[62952]: _type = "Task" [ 706.579182] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.589084] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263090, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.595957] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Releasing lock "refresh_cache-11139a80-26b2-4bc6-af91-18cd3adbe8fe" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.596125] env[62952]: DEBUG nova.compute.manager [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 706.596310] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 706.598954] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ecbb6ef7-46af-48e1-9db1-d3394ad88451 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.607026] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a84eb8-bb07-425a-9de9-a3b4116e2456 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.632074] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 11139a80-26b2-4bc6-af91-18cd3adbe8fe could not be found. [ 706.632323] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 706.632496] env[62952]: INFO nova.compute.manager [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Took 0.04 seconds to destroy the instance on the hypervisor. [ 706.632729] env[62952]: DEBUG oslo.service.loopingcall [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 706.635543] env[62952]: DEBUG nova.compute.manager [-] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 706.635649] env[62952]: DEBUG nova.network.neutron [-] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 706.656331] env[62952]: DEBUG nova.compute.manager [None req-679b48a4-523e-4524-bf18-6bcf7b312fc1 tempest-ServerDiagnosticsV248Test-1256768426 tempest-ServerDiagnosticsV248Test-1256768426-project-admin] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 706.657541] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b08999-6985-428a-b364-226f1f625744 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.661519] env[62952]: DEBUG nova.network.neutron [-] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 706.669099] env[62952]: INFO nova.compute.manager [None req-679b48a4-523e-4524-bf18-6bcf7b312fc1 tempest-ServerDiagnosticsV248Test-1256768426 tempest-ServerDiagnosticsV248Test-1256768426-project-admin] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Retrieving diagnostics [ 706.669899] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b29009-67ad-425d-a011-b7fffa2ce00c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.735428] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 706.747431] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb34d48a-e4fc-4a6c-b724-0e0b9482e105 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.755019] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c603e261-d8d2-4fbb-8140-79434462b03f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.786934] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Successfully created port: e860ee81-765c-4885-979c-9d8c922d8a80 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 706.789295] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8d6a3a-fb91-46d9-a80c-f2a7cc85c2a9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.797243] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6d540c-1e62-41f0-9f3e-757df627f5b5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.810828] env[62952]: DEBUG nova.compute.provider_tree [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 706.899536] env[62952]: DEBUG nova.compute.manager [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 707.085967] env[62952]: DEBUG nova.compute.manager [req-5368d42a-9c79-499d-b030-e85d08aa780d req-bf45cda2-c9dd-4548-a7b6-39d3c61dc19f service nova] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Received event network-changed-4c5b2393-f9f2-4e98-8c3e-9e834192ee0f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 707.085967] env[62952]: DEBUG nova.compute.manager [req-5368d42a-9c79-499d-b030-e85d08aa780d req-bf45cda2-c9dd-4548-a7b6-39d3c61dc19f service nova] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Refreshing instance network info cache due to event network-changed-4c5b2393-f9f2-4e98-8c3e-9e834192ee0f. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 707.085967] env[62952]: DEBUG oslo_concurrency.lockutils [req-5368d42a-9c79-499d-b030-e85d08aa780d req-bf45cda2-c9dd-4548-a7b6-39d3c61dc19f service nova] Acquiring lock "refresh_cache-11139a80-26b2-4bc6-af91-18cd3adbe8fe" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.085967] env[62952]: DEBUG oslo_concurrency.lockutils [req-5368d42a-9c79-499d-b030-e85d08aa780d req-bf45cda2-c9dd-4548-a7b6-39d3c61dc19f service nova] Acquired lock "refresh_cache-11139a80-26b2-4bc6-af91-18cd3adbe8fe" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.085967] env[62952]: DEBUG nova.network.neutron [req-5368d42a-9c79-499d-b030-e85d08aa780d req-bf45cda2-c9dd-4548-a7b6-39d3c61dc19f service nova] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Refreshing network info cache for port 4c5b2393-f9f2-4e98-8c3e-9e834192ee0f {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 707.092227] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263090, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.120888} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.092399] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 707.093030] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-898b8094-c656-41a7-9d30-74dca8cc19ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.116300] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Reconfiguring VM instance instance-00000015 to attach disk [datastore2] f075161d-7447-4601-a06c-514fc4dd2eb9/f075161d-7447-4601-a06c-514fc4dd2eb9.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 707.116957] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7be59340-53fb-4ac3-9c6f-c7765dbb29f4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.139692] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 707.139692] env[62952]: value = "task-1263091" [ 707.139692] env[62952]: _type = "Task" [ 707.139692] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.148620] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263091, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.168862] env[62952]: DEBUG nova.network.neutron [-] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.343112] env[62952]: ERROR nova.scheduler.client.report [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [req-60a3565d-394b-484c-966b-cea9954e2634] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 615127b5-dc82-4491-9c8f-4e6d0caa0690. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-60a3565d-394b-484c-966b-cea9954e2634"}]} [ 707.369711] env[62952]: DEBUG nova.scheduler.client.report [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Refreshing inventories for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 707.390075] env[62952]: DEBUG nova.scheduler.client.report [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Updating ProviderTree inventory for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 707.390212] env[62952]: DEBUG nova.compute.provider_tree [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 707.405751] env[62952]: DEBUG nova.scheduler.client.report [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Refreshing aggregate associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 707.426049] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.438308] env[62952]: DEBUG nova.scheduler.client.report [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Refreshing trait associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 707.613308] env[62952]: DEBUG nova.network.neutron [req-5368d42a-9c79-499d-b030-e85d08aa780d req-bf45cda2-c9dd-4548-a7b6-39d3c61dc19f service nova] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.650489] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263091, 'name': ReconfigVM_Task, 'duration_secs': 0.32605} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.653061] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Reconfigured VM instance instance-00000015 to attach disk [datastore2] f075161d-7447-4601-a06c-514fc4dd2eb9/f075161d-7447-4601-a06c-514fc4dd2eb9.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 707.653922] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5af235c6-1378-42c8-96ef-8cca89aa25fb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.661105] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 707.661105] env[62952]: value = "task-1263092" [ 707.661105] env[62952]: _type = "Task" [ 707.661105] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.669998] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263092, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.672805] env[62952]: INFO nova.compute.manager [-] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Took 1.04 seconds to deallocate network for instance. [ 707.675341] env[62952]: DEBUG nova.compute.claims [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 707.675550] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.689203] env[62952]: DEBUG nova.network.neutron [req-5368d42a-9c79-499d-b030-e85d08aa780d req-bf45cda2-c9dd-4548-a7b6-39d3c61dc19f service nova] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.745511] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 707.785026] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 707.785313] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 707.785482] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 707.785671] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 707.785850] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 707.785947] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 707.786174] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 707.786329] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 707.786488] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 707.786642] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 707.786820] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 707.790874] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9083f4d5-e2c6-467a-a22b-2db75a2b71a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.801599] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec66801-64bc-4975-930d-8dddb4aa5592 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.889440] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9771dcab-7944-474d-91f2-717058e8ced4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.897715] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15772b54-3c57-497c-b539-f2252782416b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.929859] env[62952]: ERROR nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e860ee81-765c-4885-979c-9d8c922d8a80, please check neutron logs for more information. [ 707.929859] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 707.929859] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.929859] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 707.929859] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 707.929859] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 707.929859] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 707.929859] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 707.929859] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.929859] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 707.929859] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.929859] env[62952]: ERROR nova.compute.manager raise self.value [ 707.929859] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 707.929859] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 707.929859] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.929859] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 707.930673] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.930673] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 707.930673] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e860ee81-765c-4885-979c-9d8c922d8a80, please check neutron logs for more information. [ 707.930673] env[62952]: ERROR nova.compute.manager [ 707.930673] env[62952]: Traceback (most recent call last): [ 707.930673] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 707.930673] env[62952]: listener.cb(fileno) [ 707.930673] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.930673] env[62952]: result = function(*args, **kwargs) [ 707.930673] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 707.930673] env[62952]: return func(*args, **kwargs) [ 707.930673] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.930673] env[62952]: raise e [ 707.930673] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.930673] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 707.930673] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 707.930673] env[62952]: created_port_ids = self._update_ports_for_instance( [ 707.930673] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 707.930673] env[62952]: with excutils.save_and_reraise_exception(): [ 707.930673] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.930673] env[62952]: self.force_reraise() [ 707.930673] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.930673] env[62952]: raise self.value [ 707.930673] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 707.930673] env[62952]: updated_port = self._update_port( [ 707.930673] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.930673] env[62952]: _ensure_no_port_binding_failure(port) [ 707.930673] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.930673] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 707.932011] env[62952]: nova.exception.PortBindingFailed: Binding failed for port e860ee81-765c-4885-979c-9d8c922d8a80, please check neutron logs for more information. [ 707.932011] env[62952]: Removing descriptor: 19 [ 707.932011] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-651cfc92-0fdc-4f37-9885-49bf88b2fca3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.934574] env[62952]: ERROR nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e860ee81-765c-4885-979c-9d8c922d8a80, please check neutron logs for more information. [ 707.934574] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Traceback (most recent call last): [ 707.934574] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 707.934574] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] yield resources [ 707.934574] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 707.934574] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] self.driver.spawn(context, instance, image_meta, [ 707.934574] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 707.934574] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.934574] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.934574] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] vm_ref = self.build_virtual_machine(instance, [ 707.934574] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.935088] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.935088] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.935088] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] for vif in network_info: [ 707.935088] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 707.935088] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] return self._sync_wrapper(fn, *args, **kwargs) [ 707.935088] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 707.935088] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] self.wait() [ 707.935088] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 707.935088] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] self[:] = self._gt.wait() [ 707.935088] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.935088] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] return self._exit_event.wait() [ 707.935088] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 707.935088] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] result = hub.switch() [ 707.935602] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 707.935602] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] return self.greenlet.switch() [ 707.935602] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.935602] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] result = function(*args, **kwargs) [ 707.935602] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 707.935602] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] return func(*args, **kwargs) [ 707.935602] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.935602] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] raise e [ 707.935602] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.935602] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] nwinfo = self.network_api.allocate_for_instance( [ 707.935602] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 707.935602] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] created_port_ids = self._update_ports_for_instance( [ 707.935602] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 707.936148] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] with excutils.save_and_reraise_exception(): [ 707.936148] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.936148] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] self.force_reraise() [ 707.936148] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.936148] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] raise self.value [ 707.936148] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 707.936148] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] updated_port = self._update_port( [ 707.936148] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.936148] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] _ensure_no_port_binding_failure(port) [ 707.936148] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.936148] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] raise exception.PortBindingFailed(port_id=port['id']) [ 707.936148] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] nova.exception.PortBindingFailed: Binding failed for port e860ee81-765c-4885-979c-9d8c922d8a80, please check neutron logs for more information. [ 707.936148] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] [ 707.936651] env[62952]: INFO nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Terminating instance [ 707.937093] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquiring lock "refresh_cache-7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.937266] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquired lock "refresh_cache-7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.937430] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 707.941643] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c2cfe5a-2d99-4358-9561-f45e4b7fcfb3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.961440] env[62952]: DEBUG nova.compute.provider_tree [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 708.170726] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263092, 'name': Rename_Task, 'duration_secs': 0.145555} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.170979] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 708.171347] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-87ee1cbb-baaf-48ec-a2bc-88e8e57c397f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.178817] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 708.178817] env[62952]: value = "task-1263093" [ 708.178817] env[62952]: _type = "Task" [ 708.178817] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.186563] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263093, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.195780] env[62952]: DEBUG oslo_concurrency.lockutils [req-5368d42a-9c79-499d-b030-e85d08aa780d req-bf45cda2-c9dd-4548-a7b6-39d3c61dc19f service nova] Releasing lock "refresh_cache-11139a80-26b2-4bc6-af91-18cd3adbe8fe" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.195972] env[62952]: DEBUG nova.compute.manager [req-5368d42a-9c79-499d-b030-e85d08aa780d req-bf45cda2-c9dd-4548-a7b6-39d3c61dc19f service nova] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Received event network-vif-deleted-4c5b2393-f9f2-4e98-8c3e-9e834192ee0f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 708.342650] env[62952]: DEBUG oslo_concurrency.lockutils [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Acquiring lock "8ac98d96-139a-4a03-94f3-178fafa88503" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.342886] env[62952]: DEBUG oslo_concurrency.lockutils [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Lock "8ac98d96-139a-4a03-94f3-178fafa88503" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.457323] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.485574] env[62952]: ERROR nova.scheduler.client.report [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] [req-37297bc9-e899-4d33-abfa-ce99fac8d997] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 615127b5-dc82-4491-9c8f-4e6d0caa0690. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-37297bc9-e899-4d33-abfa-ce99fac8d997"}]} [ 708.504019] env[62952]: DEBUG nova.scheduler.client.report [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Refreshing inventories for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 708.519700] env[62952]: DEBUG nova.scheduler.client.report [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Updating ProviderTree inventory for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 708.519930] env[62952]: DEBUG nova.compute.provider_tree [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 708.536309] env[62952]: DEBUG nova.scheduler.client.report [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Refreshing aggregate associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 708.539154] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.566067] env[62952]: DEBUG nova.scheduler.client.report [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Refreshing trait associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 708.691372] env[62952]: DEBUG oslo_vmware.api [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263093, 'name': PowerOnVM_Task, 'duration_secs': 0.449178} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.691693] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 708.691930] env[62952]: DEBUG nova.compute.manager [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 708.692854] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e06ba159-aa4d-4971-8189-098825a06039 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.963672] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b14a1aab-dac2-4f4c-921e-6d5c58eb9605 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.971721] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468367dc-7142-401e-a430-d89cea987b1d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.002314] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7839cc0-e845-439d-9214-f4c79ba0fafb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.010161] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfe365fb-5daa-4625-9fe8-ee76a83625ea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.023494] env[62952]: DEBUG nova.compute.provider_tree [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 709.043462] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Releasing lock "refresh_cache-7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.043871] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 709.044522] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 709.044522] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91e1646b-abaf-4e65-8ece-cbed33d6cd45 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.052704] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f2a737f-9c94-4f74-90e2-7f536d4f789f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.074684] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b could not be found. [ 709.074931] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 709.075174] env[62952]: INFO nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 709.075374] env[62952]: DEBUG oslo.service.loopingcall [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 709.075602] env[62952]: DEBUG nova.compute.manager [-] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 709.075806] env[62952]: DEBUG nova.network.neutron [-] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 709.095613] env[62952]: DEBUG nova.network.neutron [-] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 709.142690] env[62952]: DEBUG nova.compute.manager [req-66d9de4e-1a0e-4a33-b138-c22a96ed6264 req-f86c4c0a-d98f-4e44-9026-e35b0d2b5b0b service nova] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Received event network-changed-e860ee81-765c-4885-979c-9d8c922d8a80 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 709.143176] env[62952]: DEBUG nova.compute.manager [req-66d9de4e-1a0e-4a33-b138-c22a96ed6264 req-f86c4c0a-d98f-4e44-9026-e35b0d2b5b0b service nova] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Refreshing instance network info cache due to event network-changed-e860ee81-765c-4885-979c-9d8c922d8a80. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 709.143472] env[62952]: DEBUG oslo_concurrency.lockutils [req-66d9de4e-1a0e-4a33-b138-c22a96ed6264 req-f86c4c0a-d98f-4e44-9026-e35b0d2b5b0b service nova] Acquiring lock "refresh_cache-7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.143589] env[62952]: DEBUG oslo_concurrency.lockutils [req-66d9de4e-1a0e-4a33-b138-c22a96ed6264 req-f86c4c0a-d98f-4e44-9026-e35b0d2b5b0b service nova] Acquired lock "refresh_cache-7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.143769] env[62952]: DEBUG nova.network.neutron [req-66d9de4e-1a0e-4a33-b138-c22a96ed6264 req-f86c4c0a-d98f-4e44-9026-e35b0d2b5b0b service nova] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Refreshing network info cache for port e860ee81-765c-4885-979c-9d8c922d8a80 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 709.214261] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.557331] env[62952]: DEBUG nova.scheduler.client.report [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Updated inventory for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with generation 54 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 709.557600] env[62952]: DEBUG nova.compute.provider_tree [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Updating resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 generation from 54 to 55 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 709.557812] env[62952]: DEBUG nova.compute.provider_tree [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 709.598394] env[62952]: DEBUG nova.network.neutron [-] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.665917] env[62952]: DEBUG nova.network.neutron [req-66d9de4e-1a0e-4a33-b138-c22a96ed6264 req-f86c4c0a-d98f-4e44-9026-e35b0d2b5b0b service nova] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 709.741936] env[62952]: DEBUG nova.network.neutron [req-66d9de4e-1a0e-4a33-b138-c22a96ed6264 req-f86c4c0a-d98f-4e44-9026-e35b0d2b5b0b service nova] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.763189] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquiring lock "f075161d-7447-4601-a06c-514fc4dd2eb9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.763443] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Lock "f075161d-7447-4601-a06c-514fc4dd2eb9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.763648] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquiring lock "f075161d-7447-4601-a06c-514fc4dd2eb9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.763825] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Lock "f075161d-7447-4601-a06c-514fc4dd2eb9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.763989] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Lock "f075161d-7447-4601-a06c-514fc4dd2eb9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.769731] env[62952]: INFO nova.compute.manager [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Terminating instance [ 709.771445] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquiring lock "refresh_cache-f075161d-7447-4601-a06c-514fc4dd2eb9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.771664] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquired lock "refresh_cache-f075161d-7447-4601-a06c-514fc4dd2eb9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.771754] env[62952]: DEBUG nova.network.neutron [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 710.064705] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 4.339s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.067288] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.622s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.068808] env[62952]: INFO nova.compute.claims [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 710.096400] env[62952]: INFO nova.scheduler.client.report [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Deleted allocations for instance dc025711-45be-49b6-a765-f7687b94e6a1 [ 710.100504] env[62952]: INFO nova.compute.manager [-] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Took 1.02 seconds to deallocate network for instance. [ 710.103393] env[62952]: DEBUG nova.compute.claims [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 710.103572] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.243727] env[62952]: DEBUG oslo_concurrency.lockutils [req-66d9de4e-1a0e-4a33-b138-c22a96ed6264 req-f86c4c0a-d98f-4e44-9026-e35b0d2b5b0b service nova] Releasing lock "refresh_cache-7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.244009] env[62952]: DEBUG nova.compute.manager [req-66d9de4e-1a0e-4a33-b138-c22a96ed6264 req-f86c4c0a-d98f-4e44-9026-e35b0d2b5b0b service nova] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Received event network-vif-deleted-e860ee81-765c-4885-979c-9d8c922d8a80 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 710.290919] env[62952]: DEBUG nova.network.neutron [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.333189] env[62952]: DEBUG nova.network.neutron [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.604353] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3dd89-b706-4e52-a391-80db08c8d1ef tempest-ServersAaction247Test-828930316 tempest-ServersAaction247Test-828930316-project-member] Lock "dc025711-45be-49b6-a765-f7687b94e6a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.069s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.836043] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Releasing lock "refresh_cache-f075161d-7447-4601-a06c-514fc4dd2eb9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.836364] env[62952]: DEBUG nova.compute.manager [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 710.836573] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 710.837565] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f75016-dd2d-4feb-bf0b-e6065bfa33f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.845850] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 710.845850] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7d0eca4-3c39-4f45-9db6-79cfb0e87425 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.852407] env[62952]: DEBUG oslo_vmware.api [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 710.852407] env[62952]: value = "task-1263094" [ 710.852407] env[62952]: _type = "Task" [ 710.852407] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.860636] env[62952]: DEBUG oslo_vmware.api [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263094, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.362098] env[62952]: DEBUG oslo_vmware.api [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263094, 'name': PowerOffVM_Task, 'duration_secs': 0.119057} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.364330] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 711.364504] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 711.364914] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-627fe379-d798-4f35-92e3-30d8df7055c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.389876] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 711.389876] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 711.389876] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Deleting the datastore file [datastore2] f075161d-7447-4601-a06c-514fc4dd2eb9 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 711.390094] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c4b8c87-775a-4ace-aab9-d8f75420e531 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.395491] env[62952]: DEBUG oslo_vmware.api [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for the task: (returnval){ [ 711.395491] env[62952]: value = "task-1263096" [ 711.395491] env[62952]: _type = "Task" [ 711.395491] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.403173] env[62952]: DEBUG oslo_vmware.api [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263096, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.415568] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c11b1f-2711-48f3-bc8c-97d4862748f3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.423135] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f6deb35-91ae-4350-b5ca-2103b6aee42e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.452279] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d524aaa-e413-40a4-95d9-101b8b61426b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.459931] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3160dc00-b2af-4968-8cc2-cd5480857d1f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.475172] env[62952]: DEBUG nova.compute.provider_tree [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.907139] env[62952]: DEBUG oslo_vmware.api [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Task: {'id': task-1263096, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.100583} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.907139] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 711.907139] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 711.907139] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 711.907139] env[62952]: INFO nova.compute.manager [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Took 1.07 seconds to destroy the instance on the hypervisor. [ 711.907386] env[62952]: DEBUG oslo.service.loopingcall [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 711.907386] env[62952]: DEBUG nova.compute.manager [-] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 711.907386] env[62952]: DEBUG nova.network.neutron [-] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 711.922476] env[62952]: DEBUG nova.network.neutron [-] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 711.977975] env[62952]: DEBUG nova.scheduler.client.report [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.425221] env[62952]: DEBUG nova.network.neutron [-] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.482942] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.416s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.483475] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 712.486058] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.109s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.487470] env[62952]: INFO nova.compute.claims [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.929100] env[62952]: INFO nova.compute.manager [-] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Took 1.02 seconds to deallocate network for instance. [ 712.992168] env[62952]: DEBUG nova.compute.utils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 712.995521] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 712.995713] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 713.036306] env[62952]: DEBUG nova.policy [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dba2a0219a8e45a49cf91e11bcf203d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b141129b4bb24edd943a45cf4304c6c0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 713.437793] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.438894] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Successfully created port: e5af67a4-3935-4376-9260-4fd7954e2d9b {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 713.496569] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 713.931851] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8611c6d4-4e8a-46c3-9206-ea852e86447c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.939735] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded7507b-35e9-497c-95ae-154b168d8b70 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.970608] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c45ba13d-3c0b-4983-a661-5b7b1c340788 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.977899] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5314f073-6203-4fdd-89e3-998e3aa2bec0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.992688] env[62952]: DEBUG nova.compute.provider_tree [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.187536] env[62952]: DEBUG nova.compute.manager [req-9a656811-6bba-4cff-bdfb-399ab72ae7b6 req-cdac1fcd-f863-4cbe-959d-4cd9cc5a238a service nova] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Received event network-changed-e5af67a4-3935-4376-9260-4fd7954e2d9b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 714.187777] env[62952]: DEBUG nova.compute.manager [req-9a656811-6bba-4cff-bdfb-399ab72ae7b6 req-cdac1fcd-f863-4cbe-959d-4cd9cc5a238a service nova] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Refreshing instance network info cache due to event network-changed-e5af67a4-3935-4376-9260-4fd7954e2d9b. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 714.188053] env[62952]: DEBUG oslo_concurrency.lockutils [req-9a656811-6bba-4cff-bdfb-399ab72ae7b6 req-cdac1fcd-f863-4cbe-959d-4cd9cc5a238a service nova] Acquiring lock "refresh_cache-36f778a3-d795-467c-92d9-97e07b04956c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.188225] env[62952]: DEBUG oslo_concurrency.lockutils [req-9a656811-6bba-4cff-bdfb-399ab72ae7b6 req-cdac1fcd-f863-4cbe-959d-4cd9cc5a238a service nova] Acquired lock "refresh_cache-36f778a3-d795-467c-92d9-97e07b04956c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.188458] env[62952]: DEBUG nova.network.neutron [req-9a656811-6bba-4cff-bdfb-399ab72ae7b6 req-cdac1fcd-f863-4cbe-959d-4cd9cc5a238a service nova] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Refreshing network info cache for port e5af67a4-3935-4376-9260-4fd7954e2d9b {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 714.354328] env[62952]: ERROR nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e5af67a4-3935-4376-9260-4fd7954e2d9b, please check neutron logs for more information. [ 714.354328] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 714.354328] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.354328] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 714.354328] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 714.354328] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 714.354328] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 714.354328] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 714.354328] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.354328] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 714.354328] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.354328] env[62952]: ERROR nova.compute.manager raise self.value [ 714.354328] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 714.354328] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 714.354328] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.354328] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 714.354818] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.354818] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 714.354818] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e5af67a4-3935-4376-9260-4fd7954e2d9b, please check neutron logs for more information. [ 714.354818] env[62952]: ERROR nova.compute.manager [ 714.354818] env[62952]: Traceback (most recent call last): [ 714.354818] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 714.354818] env[62952]: listener.cb(fileno) [ 714.354818] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.354818] env[62952]: result = function(*args, **kwargs) [ 714.354818] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 714.354818] env[62952]: return func(*args, **kwargs) [ 714.354818] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.354818] env[62952]: raise e [ 714.354818] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.354818] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 714.354818] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 714.354818] env[62952]: created_port_ids = self._update_ports_for_instance( [ 714.354818] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 714.354818] env[62952]: with excutils.save_and_reraise_exception(): [ 714.354818] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.354818] env[62952]: self.force_reraise() [ 714.354818] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.354818] env[62952]: raise self.value [ 714.354818] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 714.354818] env[62952]: updated_port = self._update_port( [ 714.354818] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.354818] env[62952]: _ensure_no_port_binding_failure(port) [ 714.354818] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.354818] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 714.355554] env[62952]: nova.exception.PortBindingFailed: Binding failed for port e5af67a4-3935-4376-9260-4fd7954e2d9b, please check neutron logs for more information. [ 714.355554] env[62952]: Removing descriptor: 19 [ 714.496018] env[62952]: DEBUG nova.scheduler.client.report [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.507063] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 714.533056] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 714.533368] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 714.533528] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 714.533706] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 714.533850] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 714.534058] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 714.534374] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 714.534596] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 714.534808] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 714.535047] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 714.535318] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 714.536186] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d6830b8-6f05-49cb-8423-9558b7597e9d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.544546] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7448a1cb-eef2-4386-9acf-29daf461795d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.557565] env[62952]: ERROR nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e5af67a4-3935-4376-9260-4fd7954e2d9b, please check neutron logs for more information. [ 714.557565] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Traceback (most recent call last): [ 714.557565] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 714.557565] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] yield resources [ 714.557565] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 714.557565] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] self.driver.spawn(context, instance, image_meta, [ 714.557565] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 714.557565] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.557565] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.557565] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] vm_ref = self.build_virtual_machine(instance, [ 714.557565] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.557923] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.557923] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.557923] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] for vif in network_info: [ 714.557923] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 714.557923] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] return self._sync_wrapper(fn, *args, **kwargs) [ 714.557923] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 714.557923] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] self.wait() [ 714.557923] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 714.557923] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] self[:] = self._gt.wait() [ 714.557923] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.557923] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] return self._exit_event.wait() [ 714.557923] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 714.557923] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] current.throw(*self._exc) [ 714.558519] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.558519] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] result = function(*args, **kwargs) [ 714.558519] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 714.558519] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] return func(*args, **kwargs) [ 714.558519] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.558519] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] raise e [ 714.558519] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.558519] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] nwinfo = self.network_api.allocate_for_instance( [ 714.558519] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 714.558519] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] created_port_ids = self._update_ports_for_instance( [ 714.558519] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 714.558519] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] with excutils.save_and_reraise_exception(): [ 714.558519] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.558868] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] self.force_reraise() [ 714.558868] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.558868] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] raise self.value [ 714.558868] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 714.558868] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] updated_port = self._update_port( [ 714.558868] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.558868] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] _ensure_no_port_binding_failure(port) [ 714.558868] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.558868] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] raise exception.PortBindingFailed(port_id=port['id']) [ 714.558868] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] nova.exception.PortBindingFailed: Binding failed for port e5af67a4-3935-4376-9260-4fd7954e2d9b, please check neutron logs for more information. [ 714.558868] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] [ 714.558868] env[62952]: INFO nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Terminating instance [ 714.559928] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquiring lock "refresh_cache-36f778a3-d795-467c-92d9-97e07b04956c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.705076] env[62952]: DEBUG nova.network.neutron [req-9a656811-6bba-4cff-bdfb-399ab72ae7b6 req-cdac1fcd-f863-4cbe-959d-4cd9cc5a238a service nova] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 714.776098] env[62952]: DEBUG nova.network.neutron [req-9a656811-6bba-4cff-bdfb-399ab72ae7b6 req-cdac1fcd-f863-4cbe-959d-4cd9cc5a238a service nova] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.999662] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.000184] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 715.002769] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.370s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.004179] env[62952]: INFO nova.compute.claims [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 715.279229] env[62952]: DEBUG oslo_concurrency.lockutils [req-9a656811-6bba-4cff-bdfb-399ab72ae7b6 req-cdac1fcd-f863-4cbe-959d-4cd9cc5a238a service nova] Releasing lock "refresh_cache-36f778a3-d795-467c-92d9-97e07b04956c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.279687] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquired lock "refresh_cache-36f778a3-d795-467c-92d9-97e07b04956c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.279853] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 715.508421] env[62952]: DEBUG nova.compute.utils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 715.511927] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 715.512113] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 715.556029] env[62952]: DEBUG nova.policy [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dba2a0219a8e45a49cf91e11bcf203d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b141129b4bb24edd943a45cf4304c6c0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 715.800041] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.862278] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Successfully created port: 7b9602b0-50f1-4711-826f-2fa2c7dea325 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 715.930306] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.014679] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 716.216959] env[62952]: DEBUG nova.compute.manager [req-251999c1-a2db-4f39-b54e-a2d1a2e1b5e1 req-169a21be-e365-4b4f-bc33-3d1f85d8955f service nova] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Received event network-vif-deleted-e5af67a4-3935-4376-9260-4fd7954e2d9b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 716.430419] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b1bd746-6d73-4e1a-b86c-c50d0abb2166 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.433306] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Releasing lock "refresh_cache-36f778a3-d795-467c-92d9-97e07b04956c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.433749] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 716.433873] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 716.434140] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-707fa34d-75cf-4bd1-868f-2d8ab2ccd802 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.441837] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-296e93ff-77e7-4d74-b09b-3756c896e881 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.447573] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea9bfda-70dc-48be-a7d5-4309d5c08194 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.493079] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d64713c3-e46e-42bb-9815-ff8579ae744c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.493079] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 36f778a3-d795-467c-92d9-97e07b04956c could not be found. [ 716.493079] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 716.493258] env[62952]: INFO nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Took 0.06 seconds to destroy the instance on the hypervisor. [ 716.493426] env[62952]: DEBUG oslo.service.loopingcall [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 716.493651] env[62952]: DEBUG nova.compute.manager [-] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 716.493745] env[62952]: DEBUG nova.network.neutron [-] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 716.502122] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70fd1494-e729-406a-944a-bcb9d1bfd859 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.515410] env[62952]: DEBUG nova.compute.provider_tree [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.517181] env[62952]: DEBUG nova.network.neutron [-] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 716.680582] env[62952]: ERROR nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7b9602b0-50f1-4711-826f-2fa2c7dea325, please check neutron logs for more information. [ 716.680582] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 716.680582] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.680582] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 716.680582] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 716.680582] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 716.680582] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 716.680582] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 716.680582] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.680582] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 716.680582] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.680582] env[62952]: ERROR nova.compute.manager raise self.value [ 716.680582] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 716.680582] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 716.680582] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.680582] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 716.681325] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.681325] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 716.681325] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7b9602b0-50f1-4711-826f-2fa2c7dea325, please check neutron logs for more information. [ 716.681325] env[62952]: ERROR nova.compute.manager [ 716.681325] env[62952]: Traceback (most recent call last): [ 716.681325] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 716.681325] env[62952]: listener.cb(fileno) [ 716.681325] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.681325] env[62952]: result = function(*args, **kwargs) [ 716.681325] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 716.681325] env[62952]: return func(*args, **kwargs) [ 716.681325] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.681325] env[62952]: raise e [ 716.681325] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.681325] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 716.681325] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 716.681325] env[62952]: created_port_ids = self._update_ports_for_instance( [ 716.681325] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 716.681325] env[62952]: with excutils.save_and_reraise_exception(): [ 716.681325] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.681325] env[62952]: self.force_reraise() [ 716.681325] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.681325] env[62952]: raise self.value [ 716.681325] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 716.681325] env[62952]: updated_port = self._update_port( [ 716.681325] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.681325] env[62952]: _ensure_no_port_binding_failure(port) [ 716.681325] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.681325] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 716.682185] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 7b9602b0-50f1-4711-826f-2fa2c7dea325, please check neutron logs for more information. [ 716.682185] env[62952]: Removing descriptor: 19 [ 717.023078] env[62952]: DEBUG nova.network.neutron [-] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.025018] env[62952]: DEBUG nova.scheduler.client.report [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 717.030337] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 717.055875] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 717.056149] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 717.056310] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 717.056491] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 717.056636] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 717.056781] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 717.057074] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 717.057268] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 717.057441] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 717.057602] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 717.057770] env[62952]: DEBUG nova.virt.hardware [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 717.058854] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa95fcd7-331c-4130-be61-e93488b266d0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.067085] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5deeb0e6-44ce-4f83-a5bb-c25c7cd724bd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.081294] env[62952]: ERROR nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7b9602b0-50f1-4711-826f-2fa2c7dea325, please check neutron logs for more information. [ 717.081294] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Traceback (most recent call last): [ 717.081294] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 717.081294] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] yield resources [ 717.081294] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 717.081294] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] self.driver.spawn(context, instance, image_meta, [ 717.081294] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 717.081294] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.081294] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.081294] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] vm_ref = self.build_virtual_machine(instance, [ 717.081294] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.081674] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.081674] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.081674] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] for vif in network_info: [ 717.081674] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 717.081674] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] return self._sync_wrapper(fn, *args, **kwargs) [ 717.081674] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 717.081674] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] self.wait() [ 717.081674] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 717.081674] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] self[:] = self._gt.wait() [ 717.081674] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.081674] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] return self._exit_event.wait() [ 717.081674] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 717.081674] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] current.throw(*self._exc) [ 717.082052] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.082052] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] result = function(*args, **kwargs) [ 717.082052] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 717.082052] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] return func(*args, **kwargs) [ 717.082052] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.082052] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] raise e [ 717.082052] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.082052] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] nwinfo = self.network_api.allocate_for_instance( [ 717.082052] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 717.082052] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] created_port_ids = self._update_ports_for_instance( [ 717.082052] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 717.082052] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] with excutils.save_and_reraise_exception(): [ 717.082052] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.082360] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] self.force_reraise() [ 717.082360] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.082360] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] raise self.value [ 717.082360] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 717.082360] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] updated_port = self._update_port( [ 717.082360] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.082360] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] _ensure_no_port_binding_failure(port) [ 717.082360] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.082360] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] raise exception.PortBindingFailed(port_id=port['id']) [ 717.082360] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] nova.exception.PortBindingFailed: Binding failed for port 7b9602b0-50f1-4711-826f-2fa2c7dea325, please check neutron logs for more information. [ 717.082360] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] [ 717.082360] env[62952]: INFO nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Terminating instance [ 717.083598] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquiring lock "refresh_cache-7419de8d-5198-4152-a410-818af9c3b8e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.083757] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquired lock "refresh_cache-7419de8d-5198-4152-a410-818af9c3b8e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.083921] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 717.226011] env[62952]: DEBUG nova.compute.manager [None req-a086a12f-9fa6-481f-b993-749dbd1cb4fb tempest-ServerDiagnosticsV248Test-1256768426 tempest-ServerDiagnosticsV248Test-1256768426-project-admin] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 717.227201] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ebe284-d4e4-4e14-a675-4f0618363517 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.233608] env[62952]: INFO nova.compute.manager [None req-a086a12f-9fa6-481f-b993-749dbd1cb4fb tempest-ServerDiagnosticsV248Test-1256768426 tempest-ServerDiagnosticsV248Test-1256768426-project-admin] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Retrieving diagnostics [ 717.234302] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f899d1-ef37-4c9e-b34d-555490743a13 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.529560] env[62952]: INFO nova.compute.manager [-] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Took 1.04 seconds to deallocate network for instance. [ 717.530206] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.527s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.530661] env[62952]: DEBUG nova.compute.manager [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 717.534367] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.891s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.537224] env[62952]: DEBUG nova.compute.claims [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 717.537404] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.602572] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.685470] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.038524] env[62952]: DEBUG nova.compute.utils [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 718.043720] env[62952]: DEBUG nova.compute.manager [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 718.043989] env[62952]: DEBUG nova.network.neutron [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 718.098716] env[62952]: DEBUG nova.policy [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c107c5a9c73c47e3b01712a352e9e05d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '919649f57b5749f59dc9caf560e966ca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 718.190435] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Releasing lock "refresh_cache-7419de8d-5198-4152-a410-818af9c3b8e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.190435] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 718.190599] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 718.190839] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-abdf150d-a537-4a56-a060-9bae7575a58e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.202149] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa901b3b-c2b6-437c-8237-389930f79355 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.225312] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7419de8d-5198-4152-a410-818af9c3b8e2 could not be found. [ 718.225548] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 718.225727] env[62952]: INFO nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 718.226634] env[62952]: DEBUG oslo.service.loopingcall [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.229243] env[62952]: DEBUG nova.compute.manager [-] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.229367] env[62952]: DEBUG nova.network.neutron [-] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 718.244322] env[62952]: DEBUG nova.compute.manager [req-6965bbe8-d5ba-4cb6-abb8-efe5d17928e7 req-fc285c62-6462-46a0-ad15-d9112485845d service nova] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Received event network-changed-7b9602b0-50f1-4711-826f-2fa2c7dea325 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 718.244503] env[62952]: DEBUG nova.compute.manager [req-6965bbe8-d5ba-4cb6-abb8-efe5d17928e7 req-fc285c62-6462-46a0-ad15-d9112485845d service nova] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Refreshing instance network info cache due to event network-changed-7b9602b0-50f1-4711-826f-2fa2c7dea325. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 718.244704] env[62952]: DEBUG oslo_concurrency.lockutils [req-6965bbe8-d5ba-4cb6-abb8-efe5d17928e7 req-fc285c62-6462-46a0-ad15-d9112485845d service nova] Acquiring lock "refresh_cache-7419de8d-5198-4152-a410-818af9c3b8e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.244839] env[62952]: DEBUG oslo_concurrency.lockutils [req-6965bbe8-d5ba-4cb6-abb8-efe5d17928e7 req-fc285c62-6462-46a0-ad15-d9112485845d service nova] Acquired lock "refresh_cache-7419de8d-5198-4152-a410-818af9c3b8e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.244988] env[62952]: DEBUG nova.network.neutron [req-6965bbe8-d5ba-4cb6-abb8-efe5d17928e7 req-fc285c62-6462-46a0-ad15-d9112485845d service nova] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Refreshing network info cache for port 7b9602b0-50f1-4711-826f-2fa2c7dea325 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 718.250667] env[62952]: DEBUG nova.network.neutron [-] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 718.376451] env[62952]: DEBUG nova.network.neutron [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Successfully created port: 141f9cd0-39d5-42f1-944e-bded6127f741 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 718.413858] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebda5f9f-0f71-4a14-8fb1-fa2116ca2544 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.422558] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa7f8911-abff-460f-85e1-f13a78f8d6ee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.451953] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8bd2d3-a7c0-47f5-85ab-d38bf5ffe761 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.459209] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7099d96a-9b84-46d4-92e5-b1815748c6f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.463584] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Acquiring lock "80a30ea8-e260-4b92-9099-be374333896f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.463584] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Lock "80a30ea8-e260-4b92-9099-be374333896f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.463584] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Acquiring lock "80a30ea8-e260-4b92-9099-be374333896f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.463741] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Lock "80a30ea8-e260-4b92-9099-be374333896f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.463857] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Lock "80a30ea8-e260-4b92-9099-be374333896f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.474361] env[62952]: DEBUG nova.compute.provider_tree [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.475878] env[62952]: INFO nova.compute.manager [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Terminating instance [ 718.479920] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Acquiring lock "refresh_cache-80a30ea8-e260-4b92-9099-be374333896f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.479920] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Acquired lock "refresh_cache-80a30ea8-e260-4b92-9099-be374333896f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.479920] env[62952]: DEBUG nova.network.neutron [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 718.548975] env[62952]: DEBUG nova.compute.manager [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 718.752847] env[62952]: DEBUG nova.network.neutron [-] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.766257] env[62952]: DEBUG nova.network.neutron [req-6965bbe8-d5ba-4cb6-abb8-efe5d17928e7 req-fc285c62-6462-46a0-ad15-d9112485845d service nova] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 718.890744] env[62952]: DEBUG nova.network.neutron [req-6965bbe8-d5ba-4cb6-abb8-efe5d17928e7 req-fc285c62-6462-46a0-ad15-d9112485845d service nova] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.980677] env[62952]: DEBUG nova.scheduler.client.report [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 719.005664] env[62952]: DEBUG nova.network.neutron [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 719.078306] env[62952]: DEBUG nova.network.neutron [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.255578] env[62952]: INFO nova.compute.manager [-] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Took 1.03 seconds to deallocate network for instance. [ 719.258265] env[62952]: DEBUG nova.compute.claims [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 719.258447] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.368636] env[62952]: ERROR nova.compute.manager [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 141f9cd0-39d5-42f1-944e-bded6127f741, please check neutron logs for more information. [ 719.368636] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 719.368636] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.368636] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 719.368636] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 719.368636] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 719.368636] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 719.368636] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 719.368636] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.368636] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 719.368636] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.368636] env[62952]: ERROR nova.compute.manager raise self.value [ 719.368636] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 719.368636] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 719.368636] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.368636] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 719.369144] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.369144] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 719.369144] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 141f9cd0-39d5-42f1-944e-bded6127f741, please check neutron logs for more information. [ 719.369144] env[62952]: ERROR nova.compute.manager [ 719.369144] env[62952]: Traceback (most recent call last): [ 719.369144] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 719.369144] env[62952]: listener.cb(fileno) [ 719.369144] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.369144] env[62952]: result = function(*args, **kwargs) [ 719.369144] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 719.369144] env[62952]: return func(*args, **kwargs) [ 719.369144] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 719.369144] env[62952]: raise e [ 719.369144] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.369144] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 719.369144] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 719.369144] env[62952]: created_port_ids = self._update_ports_for_instance( [ 719.369144] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 719.369144] env[62952]: with excutils.save_and_reraise_exception(): [ 719.369144] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.369144] env[62952]: self.force_reraise() [ 719.369144] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.369144] env[62952]: raise self.value [ 719.369144] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 719.369144] env[62952]: updated_port = self._update_port( [ 719.369144] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.369144] env[62952]: _ensure_no_port_binding_failure(port) [ 719.369144] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.369144] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 719.369950] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 141f9cd0-39d5-42f1-944e-bded6127f741, please check neutron logs for more information. [ 719.369950] env[62952]: Removing descriptor: 19 [ 719.392597] env[62952]: DEBUG oslo_concurrency.lockutils [req-6965bbe8-d5ba-4cb6-abb8-efe5d17928e7 req-fc285c62-6462-46a0-ad15-d9112485845d service nova] Releasing lock "refresh_cache-7419de8d-5198-4152-a410-818af9c3b8e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.393098] env[62952]: DEBUG nova.compute.manager [req-6965bbe8-d5ba-4cb6-abb8-efe5d17928e7 req-fc285c62-6462-46a0-ad15-d9112485845d service nova] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Received event network-vif-deleted-7b9602b0-50f1-4711-826f-2fa2c7dea325 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 719.489230] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.955s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.490127] env[62952]: ERROR nova.compute.manager [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2e9ca01d-0d4c-484c-a227-21069ad4c486, please check neutron logs for more information. [ 719.490127] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] Traceback (most recent call last): [ 719.490127] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 719.490127] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] self.driver.spawn(context, instance, image_meta, [ 719.490127] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 719.490127] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] self._vmops.spawn(context, instance, image_meta, injected_files, [ 719.490127] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 719.490127] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] vm_ref = self.build_virtual_machine(instance, [ 719.490127] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 719.490127] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] vif_infos = vmwarevif.get_vif_info(self._session, [ 719.490127] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 719.490479] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] for vif in network_info: [ 719.490479] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 719.490479] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] return self._sync_wrapper(fn, *args, **kwargs) [ 719.490479] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 719.490479] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] self.wait() [ 719.490479] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 719.490479] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] self[:] = self._gt.wait() [ 719.490479] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 719.490479] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] return self._exit_event.wait() [ 719.490479] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 719.490479] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] result = hub.switch() [ 719.490479] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 719.490479] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] return self.greenlet.switch() [ 719.490788] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.490788] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] result = function(*args, **kwargs) [ 719.490788] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 719.490788] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] return func(*args, **kwargs) [ 719.490788] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 719.490788] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] raise e [ 719.490788] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.490788] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] nwinfo = self.network_api.allocate_for_instance( [ 719.490788] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 719.490788] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] created_port_ids = self._update_ports_for_instance( [ 719.490788] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 719.490788] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] with excutils.save_and_reraise_exception(): [ 719.490788] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.491100] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] self.force_reraise() [ 719.491100] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.491100] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] raise self.value [ 719.491100] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 719.491100] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] updated_port = self._update_port( [ 719.491100] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.491100] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] _ensure_no_port_binding_failure(port) [ 719.491100] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.491100] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] raise exception.PortBindingFailed(port_id=port['id']) [ 719.491100] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] nova.exception.PortBindingFailed: Binding failed for port 2e9ca01d-0d4c-484c-a227-21069ad4c486, please check neutron logs for more information. [ 719.491100] env[62952]: ERROR nova.compute.manager [instance: d3559857-e983-4790-8283-05aa02437723] [ 719.491361] env[62952]: DEBUG nova.compute.utils [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Binding failed for port 2e9ca01d-0d4c-484c-a227-21069ad4c486, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 719.492351] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.430s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.493958] env[62952]: INFO nova.compute.claims [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 719.497258] env[62952]: DEBUG nova.compute.manager [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Build of instance d3559857-e983-4790-8283-05aa02437723 was re-scheduled: Binding failed for port 2e9ca01d-0d4c-484c-a227-21069ad4c486, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 719.497705] env[62952]: DEBUG nova.compute.manager [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 719.497919] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Acquiring lock "refresh_cache-d3559857-e983-4790-8283-05aa02437723" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.498073] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Acquired lock "refresh_cache-d3559857-e983-4790-8283-05aa02437723" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.498299] env[62952]: DEBUG nova.network.neutron [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 719.558833] env[62952]: DEBUG nova.compute.manager [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 719.582757] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Releasing lock "refresh_cache-80a30ea8-e260-4b92-9099-be374333896f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.583175] env[62952]: DEBUG nova.compute.manager [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 719.583362] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 719.585499] env[62952]: DEBUG nova.virt.hardware [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 719.585722] env[62952]: DEBUG nova.virt.hardware [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 719.585875] env[62952]: DEBUG nova.virt.hardware [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 719.586065] env[62952]: DEBUG nova.virt.hardware [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 719.586211] env[62952]: DEBUG nova.virt.hardware [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 719.586354] env[62952]: DEBUG nova.virt.hardware [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 719.586553] env[62952]: DEBUG nova.virt.hardware [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 719.586705] env[62952]: DEBUG nova.virt.hardware [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 719.586865] env[62952]: DEBUG nova.virt.hardware [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 719.587060] env[62952]: DEBUG nova.virt.hardware [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 719.587250] env[62952]: DEBUG nova.virt.hardware [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 719.588111] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6967033e-9a94-4b46-afc5-3936defae17f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.591230] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c824f572-6670-434a-aa23-057e0763479e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.600655] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4531f9-ca70-4ac1-b554-6693a79d73b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.604595] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 719.604595] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-22ffc6ae-365d-45b9-809c-07f3cb0caf3f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.616097] env[62952]: ERROR nova.compute.manager [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 141f9cd0-39d5-42f1-944e-bded6127f741, please check neutron logs for more information. [ 719.616097] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Traceback (most recent call last): [ 719.616097] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 719.616097] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] yield resources [ 719.616097] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 719.616097] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] self.driver.spawn(context, instance, image_meta, [ 719.616097] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 719.616097] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] self._vmops.spawn(context, instance, image_meta, injected_files, [ 719.616097] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 719.616097] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] vm_ref = self.build_virtual_machine(instance, [ 719.616097] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 719.616460] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] vif_infos = vmwarevif.get_vif_info(self._session, [ 719.616460] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 719.616460] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] for vif in network_info: [ 719.616460] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 719.616460] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] return self._sync_wrapper(fn, *args, **kwargs) [ 719.616460] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 719.616460] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] self.wait() [ 719.616460] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 719.616460] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] self[:] = self._gt.wait() [ 719.616460] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 719.616460] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] return self._exit_event.wait() [ 719.616460] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 719.616460] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] current.throw(*self._exc) [ 719.616781] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.616781] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] result = function(*args, **kwargs) [ 719.616781] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 719.616781] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] return func(*args, **kwargs) [ 719.616781] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 719.616781] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] raise e [ 719.616781] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.616781] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] nwinfo = self.network_api.allocate_for_instance( [ 719.616781] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 719.616781] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] created_port_ids = self._update_ports_for_instance( [ 719.616781] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 719.616781] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] with excutils.save_and_reraise_exception(): [ 719.616781] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.617113] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] self.force_reraise() [ 719.617113] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.617113] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] raise self.value [ 719.617113] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 719.617113] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] updated_port = self._update_port( [ 719.617113] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.617113] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] _ensure_no_port_binding_failure(port) [ 719.617113] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.617113] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] raise exception.PortBindingFailed(port_id=port['id']) [ 719.617113] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] nova.exception.PortBindingFailed: Binding failed for port 141f9cd0-39d5-42f1-944e-bded6127f741, please check neutron logs for more information. [ 719.617113] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] [ 719.617113] env[62952]: INFO nova.compute.manager [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Terminating instance [ 719.619055] env[62952]: DEBUG oslo_vmware.api [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Waiting for the task: (returnval){ [ 719.619055] env[62952]: value = "task-1263097" [ 719.619055] env[62952]: _type = "Task" [ 719.619055] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.619522] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Acquiring lock "refresh_cache-0cdbad9a-9d52-45b0-8267-e181a5d38134" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.619678] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Acquired lock "refresh_cache-0cdbad9a-9d52-45b0-8267-e181a5d38134" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.619841] env[62952]: DEBUG nova.network.neutron [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 719.627881] env[62952]: DEBUG oslo_vmware.api [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': task-1263097, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.017237] env[62952]: DEBUG nova.network.neutron [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 720.093947] env[62952]: DEBUG nova.network.neutron [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.131772] env[62952]: DEBUG oslo_vmware.api [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': task-1263097, 'name': PowerOffVM_Task, 'duration_secs': 0.115683} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.132056] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 720.132231] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 720.132471] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e4f59593-18f7-4648-a905-38353a5ef396 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.139935] env[62952]: DEBUG nova.network.neutron [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 720.155635] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 720.155859] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 720.156057] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Deleting the datastore file [datastore1] 80a30ea8-e260-4b92-9099-be374333896f {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 720.156314] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1246c38b-45e1-4d7c-8e1f-9e6d62baf890 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.162473] env[62952]: DEBUG oslo_vmware.api [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Waiting for the task: (returnval){ [ 720.162473] env[62952]: value = "task-1263099" [ 720.162473] env[62952]: _type = "Task" [ 720.162473] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.170909] env[62952]: DEBUG oslo_vmware.api [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': task-1263099, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.214793] env[62952]: DEBUG nova.network.neutron [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.271846] env[62952]: DEBUG nova.compute.manager [req-5377ca65-a9a5-485b-9a36-6df6dde3dc1e req-c3a6ac30-e137-4b28-a67d-f8b91f3a9bfc service nova] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Received event network-changed-141f9cd0-39d5-42f1-944e-bded6127f741 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 720.271846] env[62952]: DEBUG nova.compute.manager [req-5377ca65-a9a5-485b-9a36-6df6dde3dc1e req-c3a6ac30-e137-4b28-a67d-f8b91f3a9bfc service nova] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Refreshing instance network info cache due to event network-changed-141f9cd0-39d5-42f1-944e-bded6127f741. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 720.271846] env[62952]: DEBUG oslo_concurrency.lockutils [req-5377ca65-a9a5-485b-9a36-6df6dde3dc1e req-c3a6ac30-e137-4b28-a67d-f8b91f3a9bfc service nova] Acquiring lock "refresh_cache-0cdbad9a-9d52-45b0-8267-e181a5d38134" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.597817] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Releasing lock "refresh_cache-d3559857-e983-4790-8283-05aa02437723" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.598329] env[62952]: DEBUG nova.compute.manager [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 720.598579] env[62952]: DEBUG nova.compute.manager [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.598825] env[62952]: DEBUG nova.network.neutron [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 720.620665] env[62952]: DEBUG nova.network.neutron [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 720.677338] env[62952]: DEBUG oslo_vmware.api [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Task: {'id': task-1263099, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094993} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.677338] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 720.677338] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 720.677338] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 720.677338] env[62952]: INFO nova.compute.manager [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Took 1.09 seconds to destroy the instance on the hypervisor. [ 720.677728] env[62952]: DEBUG oslo.service.loopingcall [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.677728] env[62952]: DEBUG nova.compute.manager [-] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.677728] env[62952]: DEBUG nova.network.neutron [-] [instance: 80a30ea8-e260-4b92-9099-be374333896f] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 720.692284] env[62952]: DEBUG nova.network.neutron [-] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 720.717311] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Releasing lock "refresh_cache-0cdbad9a-9d52-45b0-8267-e181a5d38134" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.717713] env[62952]: DEBUG nova.compute.manager [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 720.717909] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 720.718650] env[62952]: DEBUG oslo_concurrency.lockutils [req-5377ca65-a9a5-485b-9a36-6df6dde3dc1e req-c3a6ac30-e137-4b28-a67d-f8b91f3a9bfc service nova] Acquired lock "refresh_cache-0cdbad9a-9d52-45b0-8267-e181a5d38134" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.718891] env[62952]: DEBUG nova.network.neutron [req-5377ca65-a9a5-485b-9a36-6df6dde3dc1e req-c3a6ac30-e137-4b28-a67d-f8b91f3a9bfc service nova] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Refreshing network info cache for port 141f9cd0-39d5-42f1-944e-bded6127f741 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 720.723206] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9ddaf86a-7c86-4a6e-a47d-8226b5b2e668 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.728840] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e676d562-b006-41b0-9a28-a94a2338a001 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.753964] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0cdbad9a-9d52-45b0-8267-e181a5d38134 could not be found. [ 720.754184] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 720.754353] env[62952]: INFO nova.compute.manager [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Took 0.04 seconds to destroy the instance on the hypervisor. [ 720.754596] env[62952]: DEBUG oslo.service.loopingcall [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.757591] env[62952]: DEBUG nova.compute.manager [-] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.757591] env[62952]: DEBUG nova.network.neutron [-] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 720.779734] env[62952]: DEBUG nova.network.neutron [-] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 720.943247] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-259b162a-11b7-46f7-97e5-f7be9044263b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.951548] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c4b4b97-af9e-4e6f-901b-e51d734db8fa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.981271] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c2dd642-d4b2-4440-816d-1fba563db58c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.988851] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fd22dc7-3d88-4c2c-bebf-5e1744b28bc8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.003251] env[62952]: DEBUG nova.compute.provider_tree [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 721.124036] env[62952]: DEBUG nova.network.neutron [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.195654] env[62952]: DEBUG nova.network.neutron [-] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.236395] env[62952]: DEBUG nova.network.neutron [req-5377ca65-a9a5-485b-9a36-6df6dde3dc1e req-c3a6ac30-e137-4b28-a67d-f8b91f3a9bfc service nova] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 721.282544] env[62952]: DEBUG nova.network.neutron [-] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.308147] env[62952]: DEBUG nova.network.neutron [req-5377ca65-a9a5-485b-9a36-6df6dde3dc1e req-c3a6ac30-e137-4b28-a67d-f8b91f3a9bfc service nova] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.534037] env[62952]: DEBUG nova.scheduler.client.report [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Updated inventory for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with generation 55 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 721.534330] env[62952]: DEBUG nova.compute.provider_tree [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Updating resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 generation from 55 to 56 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 721.534510] env[62952]: DEBUG nova.compute.provider_tree [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 721.627163] env[62952]: INFO nova.compute.manager [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] [instance: d3559857-e983-4790-8283-05aa02437723] Took 1.03 seconds to deallocate network for instance. [ 721.698569] env[62952]: INFO nova.compute.manager [-] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Took 1.02 seconds to deallocate network for instance. [ 721.786055] env[62952]: INFO nova.compute.manager [-] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Took 1.03 seconds to deallocate network for instance. [ 721.787746] env[62952]: DEBUG nova.compute.claims [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 721.787932] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.810952] env[62952]: DEBUG oslo_concurrency.lockutils [req-5377ca65-a9a5-485b-9a36-6df6dde3dc1e req-c3a6ac30-e137-4b28-a67d-f8b91f3a9bfc service nova] Releasing lock "refresh_cache-0cdbad9a-9d52-45b0-8267-e181a5d38134" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.811269] env[62952]: DEBUG nova.compute.manager [req-5377ca65-a9a5-485b-9a36-6df6dde3dc1e req-c3a6ac30-e137-4b28-a67d-f8b91f3a9bfc service nova] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Received event network-vif-deleted-141f9cd0-39d5-42f1-944e-bded6127f741 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 722.039651] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.547s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.040226] env[62952]: DEBUG nova.compute.manager [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 722.043587] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 20.412s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.044511] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.044748] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 722.045112] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.828s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.046594] env[62952]: INFO nova.compute.claims [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 722.049879] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42cc3fc9-9a0d-4107-ab72-037c524a6102 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.058601] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b13cf4-48b9-488b-b578-bd5a950abd72 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.073186] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a12d660-208c-44df-af9e-fa592e059812 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.080065] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ba1669-4633-4251-9348-979a621c6cde {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.111030] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181529MB free_disk=145GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 722.111030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.204844] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.550805] env[62952]: DEBUG nova.compute.utils [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 722.552196] env[62952]: DEBUG nova.compute.manager [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 722.552367] env[62952]: DEBUG nova.network.neutron [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 722.591746] env[62952]: DEBUG nova.policy [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '89a3750831f749cabb4bd38d238b70e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3154e9ad8477403ea9583387c6aabb1f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 722.653221] env[62952]: INFO nova.scheduler.client.report [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Deleted allocations for instance d3559857-e983-4790-8283-05aa02437723 [ 722.920853] env[62952]: DEBUG nova.network.neutron [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Successfully created port: 52ae2518-4f0c-4959-9d2c-00f38c4749fb {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 723.055470] env[62952]: DEBUG nova.compute.manager [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 723.162431] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79cbb25d-a23e-4aa2-ab48-bef81cb42641 tempest-ServerPasswordTestJSON-1384833766 tempest-ServerPasswordTestJSON-1384833766-project-member] Lock "d3559857-e983-4790-8283-05aa02437723" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.146s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.409862] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd580613-4d99-4480-a939-74bcd9e011cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.418908] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-035e9693-f3b6-4150-8a46-1996971df21d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.452111] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bbf4555-12c6-458b-aaba-dcc6fedafa24 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.459809] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe7ab33-0a52-457f-a757-30f96841c765 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.474000] env[62952]: DEBUG nova.compute.provider_tree [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.664176] env[62952]: DEBUG nova.compute.manager [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 723.694569] env[62952]: DEBUG nova.compute.manager [req-8df07b16-94ed-4b52-a007-070be3eba4c4 req-de1e5bfa-2d32-4b0e-8fc9-55e4ca5d209a service nova] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Received event network-changed-52ae2518-4f0c-4959-9d2c-00f38c4749fb {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 723.694741] env[62952]: DEBUG nova.compute.manager [req-8df07b16-94ed-4b52-a007-070be3eba4c4 req-de1e5bfa-2d32-4b0e-8fc9-55e4ca5d209a service nova] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Refreshing instance network info cache due to event network-changed-52ae2518-4f0c-4959-9d2c-00f38c4749fb. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 723.694905] env[62952]: DEBUG oslo_concurrency.lockutils [req-8df07b16-94ed-4b52-a007-070be3eba4c4 req-de1e5bfa-2d32-4b0e-8fc9-55e4ca5d209a service nova] Acquiring lock "refresh_cache-4bc891c1-e7e1-4196-9770-9bb42ed7499c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.695082] env[62952]: DEBUG oslo_concurrency.lockutils [req-8df07b16-94ed-4b52-a007-070be3eba4c4 req-de1e5bfa-2d32-4b0e-8fc9-55e4ca5d209a service nova] Acquired lock "refresh_cache-4bc891c1-e7e1-4196-9770-9bb42ed7499c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.695244] env[62952]: DEBUG nova.network.neutron [req-8df07b16-94ed-4b52-a007-070be3eba4c4 req-de1e5bfa-2d32-4b0e-8fc9-55e4ca5d209a service nova] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Refreshing network info cache for port 52ae2518-4f0c-4959-9d2c-00f38c4749fb {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 723.790081] env[62952]: ERROR nova.compute.manager [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 52ae2518-4f0c-4959-9d2c-00f38c4749fb, please check neutron logs for more information. [ 723.790081] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 723.790081] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.790081] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 723.790081] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 723.790081] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 723.790081] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 723.790081] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 723.790081] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.790081] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 723.790081] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.790081] env[62952]: ERROR nova.compute.manager raise self.value [ 723.790081] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 723.790081] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 723.790081] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.790081] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 723.790578] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.790578] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 723.790578] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 52ae2518-4f0c-4959-9d2c-00f38c4749fb, please check neutron logs for more information. [ 723.790578] env[62952]: ERROR nova.compute.manager [ 723.790578] env[62952]: Traceback (most recent call last): [ 723.790578] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 723.790578] env[62952]: listener.cb(fileno) [ 723.790578] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.790578] env[62952]: result = function(*args, **kwargs) [ 723.790578] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.790578] env[62952]: return func(*args, **kwargs) [ 723.790578] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.790578] env[62952]: raise e [ 723.790578] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.790578] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 723.790578] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 723.790578] env[62952]: created_port_ids = self._update_ports_for_instance( [ 723.790578] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 723.790578] env[62952]: with excutils.save_and_reraise_exception(): [ 723.790578] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.790578] env[62952]: self.force_reraise() [ 723.790578] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.790578] env[62952]: raise self.value [ 723.790578] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 723.790578] env[62952]: updated_port = self._update_port( [ 723.790578] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.790578] env[62952]: _ensure_no_port_binding_failure(port) [ 723.790578] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.790578] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 723.791378] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 52ae2518-4f0c-4959-9d2c-00f38c4749fb, please check neutron logs for more information. [ 723.791378] env[62952]: Removing descriptor: 19 [ 723.979469] env[62952]: DEBUG nova.scheduler.client.report [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 724.067443] env[62952]: DEBUG nova.compute.manager [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 724.093704] env[62952]: DEBUG nova.virt.hardware [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 724.093965] env[62952]: DEBUG nova.virt.hardware [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 724.094135] env[62952]: DEBUG nova.virt.hardware [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 724.094318] env[62952]: DEBUG nova.virt.hardware [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 724.094462] env[62952]: DEBUG nova.virt.hardware [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 724.094608] env[62952]: DEBUG nova.virt.hardware [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 724.095151] env[62952]: DEBUG nova.virt.hardware [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 724.095345] env[62952]: DEBUG nova.virt.hardware [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 724.095520] env[62952]: DEBUG nova.virt.hardware [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 724.095684] env[62952]: DEBUG nova.virt.hardware [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 724.096172] env[62952]: DEBUG nova.virt.hardware [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 724.097081] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d93b209-95e1-4322-9e99-5e6cf7b94847 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.106186] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6948e0b1-9923-483b-ab84-e8df111746e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.120758] env[62952]: ERROR nova.compute.manager [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 52ae2518-4f0c-4959-9d2c-00f38c4749fb, please check neutron logs for more information. [ 724.120758] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Traceback (most recent call last): [ 724.120758] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 724.120758] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] yield resources [ 724.120758] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 724.120758] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] self.driver.spawn(context, instance, image_meta, [ 724.120758] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 724.120758] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.120758] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.120758] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] vm_ref = self.build_virtual_machine(instance, [ 724.120758] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.121117] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.121117] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.121117] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] for vif in network_info: [ 724.121117] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.121117] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] return self._sync_wrapper(fn, *args, **kwargs) [ 724.121117] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.121117] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] self.wait() [ 724.121117] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.121117] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] self[:] = self._gt.wait() [ 724.121117] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.121117] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] return self._exit_event.wait() [ 724.121117] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 724.121117] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] current.throw(*self._exc) [ 724.121416] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.121416] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] result = function(*args, **kwargs) [ 724.121416] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 724.121416] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] return func(*args, **kwargs) [ 724.121416] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.121416] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] raise e [ 724.121416] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.121416] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] nwinfo = self.network_api.allocate_for_instance( [ 724.121416] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 724.121416] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] created_port_ids = self._update_ports_for_instance( [ 724.121416] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 724.121416] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] with excutils.save_and_reraise_exception(): [ 724.121416] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.121787] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] self.force_reraise() [ 724.121787] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.121787] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] raise self.value [ 724.121787] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 724.121787] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] updated_port = self._update_port( [ 724.121787] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.121787] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] _ensure_no_port_binding_failure(port) [ 724.121787] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.121787] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] raise exception.PortBindingFailed(port_id=port['id']) [ 724.121787] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] nova.exception.PortBindingFailed: Binding failed for port 52ae2518-4f0c-4959-9d2c-00f38c4749fb, please check neutron logs for more information. [ 724.121787] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] [ 724.121787] env[62952]: INFO nova.compute.manager [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Terminating instance [ 724.123231] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Acquiring lock "refresh_cache-4bc891c1-e7e1-4196-9770-9bb42ed7499c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.195105] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.214358] env[62952]: DEBUG nova.network.neutron [req-8df07b16-94ed-4b52-a007-070be3eba4c4 req-de1e5bfa-2d32-4b0e-8fc9-55e4ca5d209a service nova] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.289352] env[62952]: DEBUG nova.network.neutron [req-8df07b16-94ed-4b52-a007-070be3eba4c4 req-de1e5bfa-2d32-4b0e-8fc9-55e4ca5d209a service nova] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.486388] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.486911] env[62952]: DEBUG nova.compute.manager [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 724.489509] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.444s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.792328] env[62952]: DEBUG oslo_concurrency.lockutils [req-8df07b16-94ed-4b52-a007-070be3eba4c4 req-de1e5bfa-2d32-4b0e-8fc9-55e4ca5d209a service nova] Releasing lock "refresh_cache-4bc891c1-e7e1-4196-9770-9bb42ed7499c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.792328] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Acquired lock "refresh_cache-4bc891c1-e7e1-4196-9770-9bb42ed7499c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.792682] env[62952]: DEBUG nova.network.neutron [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 725.001242] env[62952]: DEBUG nova.compute.utils [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 725.002871] env[62952]: DEBUG nova.compute.manager [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 725.002970] env[62952]: DEBUG nova.network.neutron [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 725.046663] env[62952]: DEBUG nova.policy [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50dfaab33dab483a81b2bd915ca98935', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4be6c03184424deba6d91682367231db', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 725.318963] env[62952]: DEBUG nova.network.neutron [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.388573] env[62952]: DEBUG nova.network.neutron [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Successfully created port: 249319dd-b412-4d3c-9b70-05b017a04e10 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 725.451921] env[62952]: DEBUG nova.network.neutron [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.456224] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c630ad2f-92fb-428c-a6dc-da46a4f0cabf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.465312] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe70c8b-e669-45a6-9d58-7d9acdc8220d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.496812] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ffbbdbc-cbbb-40c7-991b-2e97f13d3dde {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.504255] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc0021b7-5727-490a-9886-b84495616ce1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.508751] env[62952]: DEBUG nova.compute.manager [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 725.521063] env[62952]: DEBUG nova.compute.provider_tree [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.775835] env[62952]: DEBUG nova.compute.manager [req-a06c2135-b804-469d-9eba-8b921035f73e req-e07bb5d3-cb8a-44c5-a811-835ae3cebf16 service nova] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Received event network-vif-deleted-52ae2518-4f0c-4959-9d2c-00f38c4749fb {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 725.960973] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Releasing lock "refresh_cache-4bc891c1-e7e1-4196-9770-9bb42ed7499c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.962491] env[62952]: DEBUG nova.compute.manager [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 725.962491] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 725.962491] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f8b407d-a94f-40df-a16e-0db3c703dda1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.971055] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089ace9f-8b45-42ea-b864-7f061053d031 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.002562] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4bc891c1-e7e1-4196-9770-9bb42ed7499c could not be found. [ 726.002814] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 726.002996] env[62952]: INFO nova.compute.manager [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 726.003256] env[62952]: DEBUG oslo.service.loopingcall [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 726.003480] env[62952]: DEBUG nova.compute.manager [-] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.003574] env[62952]: DEBUG nova.network.neutron [-] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 726.021246] env[62952]: DEBUG nova.network.neutron [-] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.026821] env[62952]: DEBUG nova.scheduler.client.report [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.525858] env[62952]: DEBUG nova.compute.manager [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 726.528770] env[62952]: DEBUG nova.network.neutron [-] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.531037] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.042s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.531640] env[62952]: ERROR nova.compute.manager [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d7878ac1-867d-4c80-9e77-e1080f980358, please check neutron logs for more information. [ 726.531640] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Traceback (most recent call last): [ 726.531640] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 726.531640] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] self.driver.spawn(context, instance, image_meta, [ 726.531640] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 726.531640] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.531640] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.531640] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] vm_ref = self.build_virtual_machine(instance, [ 726.531640] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.531640] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.531640] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.531920] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] for vif in network_info: [ 726.531920] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.531920] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] return self._sync_wrapper(fn, *args, **kwargs) [ 726.531920] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.531920] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] self.wait() [ 726.531920] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.531920] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] self[:] = self._gt.wait() [ 726.531920] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.531920] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] return self._exit_event.wait() [ 726.531920] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 726.531920] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] result = hub.switch() [ 726.531920] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 726.531920] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] return self.greenlet.switch() [ 726.532246] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.532246] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] result = function(*args, **kwargs) [ 726.532246] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.532246] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] return func(*args, **kwargs) [ 726.532246] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.532246] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] raise e [ 726.532246] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.532246] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] nwinfo = self.network_api.allocate_for_instance( [ 726.532246] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 726.532246] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] created_port_ids = self._update_ports_for_instance( [ 726.532246] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 726.532246] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] with excutils.save_and_reraise_exception(): [ 726.532246] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.532584] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] self.force_reraise() [ 726.532584] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.532584] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] raise self.value [ 726.532584] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 726.532584] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] updated_port = self._update_port( [ 726.532584] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.532584] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] _ensure_no_port_binding_failure(port) [ 726.532584] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.532584] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] raise exception.PortBindingFailed(port_id=port['id']) [ 726.532584] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] nova.exception.PortBindingFailed: Binding failed for port d7878ac1-867d-4c80-9e77-e1080f980358, please check neutron logs for more information. [ 726.532584] env[62952]: ERROR nova.compute.manager [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] [ 726.532844] env[62952]: DEBUG nova.compute.utils [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Binding failed for port d7878ac1-867d-4c80-9e77-e1080f980358, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 726.535386] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.108s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.538571] env[62952]: INFO nova.compute.claims [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 726.539544] env[62952]: DEBUG nova.compute.manager [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Build of instance 3516272a-9ebd-413e-9b88-1a5c83b50f61 was re-scheduled: Binding failed for port d7878ac1-867d-4c80-9e77-e1080f980358, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 726.540072] env[62952]: DEBUG nova.compute.manager [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 726.540387] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "refresh_cache-3516272a-9ebd-413e-9b88-1a5c83b50f61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.540563] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquired lock "refresh_cache-3516272a-9ebd-413e-9b88-1a5c83b50f61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.540724] env[62952]: DEBUG nova.network.neutron [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 726.564153] env[62952]: DEBUG nova.virt.hardware [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 726.564401] env[62952]: DEBUG nova.virt.hardware [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 726.564555] env[62952]: DEBUG nova.virt.hardware [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 726.564918] env[62952]: DEBUG nova.virt.hardware [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 726.565108] env[62952]: DEBUG nova.virt.hardware [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 726.565440] env[62952]: DEBUG nova.virt.hardware [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 726.566646] env[62952]: DEBUG nova.virt.hardware [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 726.566918] env[62952]: DEBUG nova.virt.hardware [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 726.567193] env[62952]: DEBUG nova.virt.hardware [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 726.567376] env[62952]: DEBUG nova.virt.hardware [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 726.567552] env[62952]: DEBUG nova.virt.hardware [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 726.568695] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a01d699-eaff-48f5-8000-e3fadcf0dc9d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.577383] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82649521-20f7-4e3b-8852-bae6a2a213c4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.675539] env[62952]: ERROR nova.compute.manager [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 249319dd-b412-4d3c-9b70-05b017a04e10, please check neutron logs for more information. [ 726.675539] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 726.675539] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.675539] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 726.675539] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 726.675539] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 726.675539] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 726.675539] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 726.675539] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.675539] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 726.675539] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.675539] env[62952]: ERROR nova.compute.manager raise self.value [ 726.675539] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 726.675539] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 726.675539] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.675539] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 726.676121] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.676121] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 726.676121] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 249319dd-b412-4d3c-9b70-05b017a04e10, please check neutron logs for more information. [ 726.676121] env[62952]: ERROR nova.compute.manager [ 726.676121] env[62952]: Traceback (most recent call last): [ 726.676121] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 726.676121] env[62952]: listener.cb(fileno) [ 726.676121] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.676121] env[62952]: result = function(*args, **kwargs) [ 726.676121] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.676121] env[62952]: return func(*args, **kwargs) [ 726.676121] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.676121] env[62952]: raise e [ 726.676121] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.676121] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 726.676121] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 726.676121] env[62952]: created_port_ids = self._update_ports_for_instance( [ 726.676121] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 726.676121] env[62952]: with excutils.save_and_reraise_exception(): [ 726.676121] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.676121] env[62952]: self.force_reraise() [ 726.676121] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.676121] env[62952]: raise self.value [ 726.676121] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 726.676121] env[62952]: updated_port = self._update_port( [ 726.676121] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.676121] env[62952]: _ensure_no_port_binding_failure(port) [ 726.676121] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.676121] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 726.676774] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 249319dd-b412-4d3c-9b70-05b017a04e10, please check neutron logs for more information. [ 726.676774] env[62952]: Removing descriptor: 19 [ 726.676774] env[62952]: ERROR nova.compute.manager [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 249319dd-b412-4d3c-9b70-05b017a04e10, please check neutron logs for more information. [ 726.676774] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Traceback (most recent call last): [ 726.676774] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 726.676774] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] yield resources [ 726.676774] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 726.676774] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] self.driver.spawn(context, instance, image_meta, [ 726.676774] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 726.676774] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.676774] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.676774] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] vm_ref = self.build_virtual_machine(instance, [ 726.677122] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.677122] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.677122] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.677122] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] for vif in network_info: [ 726.677122] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.677122] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] return self._sync_wrapper(fn, *args, **kwargs) [ 726.677122] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.677122] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] self.wait() [ 726.677122] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.677122] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] self[:] = self._gt.wait() [ 726.677122] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.677122] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] return self._exit_event.wait() [ 726.677122] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 726.677434] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] result = hub.switch() [ 726.677434] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 726.677434] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] return self.greenlet.switch() [ 726.677434] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.677434] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] result = function(*args, **kwargs) [ 726.677434] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.677434] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] return func(*args, **kwargs) [ 726.677434] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.677434] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] raise e [ 726.677434] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.677434] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] nwinfo = self.network_api.allocate_for_instance( [ 726.677434] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 726.677434] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] created_port_ids = self._update_ports_for_instance( [ 726.677763] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 726.677763] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] with excutils.save_and_reraise_exception(): [ 726.677763] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.677763] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] self.force_reraise() [ 726.677763] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.677763] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] raise self.value [ 726.677763] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 726.677763] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] updated_port = self._update_port( [ 726.677763] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.677763] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] _ensure_no_port_binding_failure(port) [ 726.677763] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.677763] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] raise exception.PortBindingFailed(port_id=port['id']) [ 726.678100] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] nova.exception.PortBindingFailed: Binding failed for port 249319dd-b412-4d3c-9b70-05b017a04e10, please check neutron logs for more information. [ 726.678100] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] [ 726.678100] env[62952]: INFO nova.compute.manager [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Terminating instance [ 726.679465] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Acquiring lock "refresh_cache-268e5249-0156-45dd-bbbd-10105d16f269" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.679724] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Acquired lock "refresh_cache-268e5249-0156-45dd-bbbd-10105d16f269" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.679870] env[62952]: DEBUG nova.network.neutron [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 727.037125] env[62952]: INFO nova.compute.manager [-] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Took 1.03 seconds to deallocate network for instance. [ 727.038293] env[62952]: DEBUG nova.compute.claims [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 727.038293] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.060385] env[62952]: DEBUG nova.network.neutron [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.130417] env[62952]: DEBUG nova.network.neutron [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.205412] env[62952]: DEBUG nova.network.neutron [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.277291] env[62952]: DEBUG nova.network.neutron [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.632634] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Releasing lock "refresh_cache-3516272a-9ebd-413e-9b88-1a5c83b50f61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.632892] env[62952]: DEBUG nova.compute.manager [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 727.633084] env[62952]: DEBUG nova.compute.manager [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 727.633261] env[62952]: DEBUG nova.network.neutron [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 727.661011] env[62952]: DEBUG nova.network.neutron [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.781709] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Releasing lock "refresh_cache-268e5249-0156-45dd-bbbd-10105d16f269" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.781894] env[62952]: DEBUG nova.compute.manager [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 727.782163] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 727.782574] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3efc520-66f5-4927-8fb9-deac6bfade80 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.793426] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9d77f70-81c0-4cd2-93d1-b7c6408aeeb6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.809901] env[62952]: DEBUG nova.compute.manager [req-b6205696-f0ee-4432-ab2c-ac5dbfac10e3 req-00891e2f-d1cb-4c4b-a9df-3bda6159dab6 service nova] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Received event network-changed-249319dd-b412-4d3c-9b70-05b017a04e10 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 727.813573] env[62952]: DEBUG nova.compute.manager [req-b6205696-f0ee-4432-ab2c-ac5dbfac10e3 req-00891e2f-d1cb-4c4b-a9df-3bda6159dab6 service nova] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Refreshing instance network info cache due to event network-changed-249319dd-b412-4d3c-9b70-05b017a04e10. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 727.813573] env[62952]: DEBUG oslo_concurrency.lockutils [req-b6205696-f0ee-4432-ab2c-ac5dbfac10e3 req-00891e2f-d1cb-4c4b-a9df-3bda6159dab6 service nova] Acquiring lock "refresh_cache-268e5249-0156-45dd-bbbd-10105d16f269" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.813573] env[62952]: DEBUG oslo_concurrency.lockutils [req-b6205696-f0ee-4432-ab2c-ac5dbfac10e3 req-00891e2f-d1cb-4c4b-a9df-3bda6159dab6 service nova] Acquired lock "refresh_cache-268e5249-0156-45dd-bbbd-10105d16f269" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.813573] env[62952]: DEBUG nova.network.neutron [req-b6205696-f0ee-4432-ab2c-ac5dbfac10e3 req-00891e2f-d1cb-4c4b-a9df-3bda6159dab6 service nova] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Refreshing network info cache for port 249319dd-b412-4d3c-9b70-05b017a04e10 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 727.819292] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 268e5249-0156-45dd-bbbd-10105d16f269 could not be found. [ 727.819450] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 727.819626] env[62952]: INFO nova.compute.manager [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Took 0.04 seconds to destroy the instance on the hypervisor. [ 727.819847] env[62952]: DEBUG oslo.service.loopingcall [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.820061] env[62952]: DEBUG nova.compute.manager [-] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 727.820151] env[62952]: DEBUG nova.network.neutron [-] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 727.844191] env[62952]: DEBUG nova.network.neutron [-] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.983453] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5df882b-a5bd-40da-bf98-71ec3488d244 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.991659] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e9b4a8c-ea18-480f-897f-0ab49b7db747 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.021626] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3629dd6f-c82e-4ebf-9706-b33b1b066eeb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.028579] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-126d8567-497b-4503-875d-8162cc0e3c20 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.046970] env[62952]: DEBUG nova.compute.provider_tree [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.163750] env[62952]: DEBUG nova.network.neutron [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.332292] env[62952]: DEBUG nova.network.neutron [req-b6205696-f0ee-4432-ab2c-ac5dbfac10e3 req-00891e2f-d1cb-4c4b-a9df-3bda6159dab6 service nova] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.348829] env[62952]: DEBUG nova.network.neutron [-] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.426855] env[62952]: DEBUG nova.network.neutron [req-b6205696-f0ee-4432-ab2c-ac5dbfac10e3 req-00891e2f-d1cb-4c4b-a9df-3bda6159dab6 service nova] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.551204] env[62952]: DEBUG nova.scheduler.client.report [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 728.667167] env[62952]: INFO nova.compute.manager [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 3516272a-9ebd-413e-9b88-1a5c83b50f61] Took 1.03 seconds to deallocate network for instance. [ 728.852121] env[62952]: INFO nova.compute.manager [-] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Took 1.03 seconds to deallocate network for instance. [ 728.857467] env[62952]: DEBUG nova.compute.claims [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 728.857467] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.929285] env[62952]: DEBUG oslo_concurrency.lockutils [req-b6205696-f0ee-4432-ab2c-ac5dbfac10e3 req-00891e2f-d1cb-4c4b-a9df-3bda6159dab6 service nova] Releasing lock "refresh_cache-268e5249-0156-45dd-bbbd-10105d16f269" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.929585] env[62952]: DEBUG nova.compute.manager [req-b6205696-f0ee-4432-ab2c-ac5dbfac10e3 req-00891e2f-d1cb-4c4b-a9df-3bda6159dab6 service nova] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Received event network-vif-deleted-249319dd-b412-4d3c-9b70-05b017a04e10 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 729.061021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.527s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.062035] env[62952]: DEBUG nova.compute.manager [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 729.064338] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.389s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.571302] env[62952]: DEBUG nova.compute.utils [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 729.573899] env[62952]: DEBUG nova.compute.manager [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 729.574265] env[62952]: DEBUG nova.network.neutron [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 729.629263] env[62952]: DEBUG nova.policy [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6967f8ebdc944dbfb2d19b5d7acdd3e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91fde3ceb94649f698459dfc174bf835', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 729.706396] env[62952]: INFO nova.scheduler.client.report [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Deleted allocations for instance 3516272a-9ebd-413e-9b88-1a5c83b50f61 [ 730.005988] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de19cce6-3f48-452b-86c5-715fa2d5ee98 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.015601] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0079ac8d-e053-4d9e-8e65-6ebfb3a0d56d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.021493] env[62952]: DEBUG nova.network.neutron [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Successfully created port: 9b472956-916e-4f25-a11f-fee191c954c1 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 730.052988] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-194376fc-6377-453d-968e-7c21ba503e41 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.060482] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-414f5c08-766b-457e-8888-8811eaf07cfb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.076268] env[62952]: DEBUG nova.compute.provider_tree [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.081020] env[62952]: DEBUG nova.compute.manager [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 730.217377] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b9ca2f0c-05eb-4e5b-b519-62eb57cb9671 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "3516272a-9ebd-413e-9b88-1a5c83b50f61" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.794s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.516711] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.516944] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.581471] env[62952]: DEBUG nova.scheduler.client.report [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.720124] env[62952]: DEBUG nova.compute.manager [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 730.858414] env[62952]: DEBUG nova.compute.manager [req-082ac7aa-1ff8-4299-883d-2caec9b8f86f req-ce5e6680-2e80-4b06-9b3e-360cf0147ee5 service nova] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Received event network-changed-9b472956-916e-4f25-a11f-fee191c954c1 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 730.858414] env[62952]: DEBUG nova.compute.manager [req-082ac7aa-1ff8-4299-883d-2caec9b8f86f req-ce5e6680-2e80-4b06-9b3e-360cf0147ee5 service nova] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Refreshing instance network info cache due to event network-changed-9b472956-916e-4f25-a11f-fee191c954c1. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 730.858414] env[62952]: DEBUG oslo_concurrency.lockutils [req-082ac7aa-1ff8-4299-883d-2caec9b8f86f req-ce5e6680-2e80-4b06-9b3e-360cf0147ee5 service nova] Acquiring lock "refresh_cache-f9ed97a4-ec32-415b-958e-d362b622b1d4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.858414] env[62952]: DEBUG oslo_concurrency.lockutils [req-082ac7aa-1ff8-4299-883d-2caec9b8f86f req-ce5e6680-2e80-4b06-9b3e-360cf0147ee5 service nova] Acquired lock "refresh_cache-f9ed97a4-ec32-415b-958e-d362b622b1d4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.858414] env[62952]: DEBUG nova.network.neutron [req-082ac7aa-1ff8-4299-883d-2caec9b8f86f req-ce5e6680-2e80-4b06-9b3e-360cf0147ee5 service nova] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Refreshing network info cache for port 9b472956-916e-4f25-a11f-fee191c954c1 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 730.937451] env[62952]: ERROR nova.compute.manager [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9b472956-916e-4f25-a11f-fee191c954c1, please check neutron logs for more information. [ 730.937451] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 730.937451] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.937451] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 730.937451] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 730.937451] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 730.937451] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 730.937451] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 730.937451] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.937451] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 730.937451] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.937451] env[62952]: ERROR nova.compute.manager raise self.value [ 730.937451] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 730.937451] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 730.937451] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.937451] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 730.937876] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.937876] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 730.937876] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9b472956-916e-4f25-a11f-fee191c954c1, please check neutron logs for more information. [ 730.937876] env[62952]: ERROR nova.compute.manager [ 730.937876] env[62952]: Traceback (most recent call last): [ 730.937876] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 730.937876] env[62952]: listener.cb(fileno) [ 730.937876] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.937876] env[62952]: result = function(*args, **kwargs) [ 730.937876] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 730.937876] env[62952]: return func(*args, **kwargs) [ 730.937876] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 730.937876] env[62952]: raise e [ 730.937876] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.937876] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 730.937876] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 730.937876] env[62952]: created_port_ids = self._update_ports_for_instance( [ 730.937876] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 730.937876] env[62952]: with excutils.save_and_reraise_exception(): [ 730.937876] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.937876] env[62952]: self.force_reraise() [ 730.937876] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.937876] env[62952]: raise self.value [ 730.937876] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 730.937876] env[62952]: updated_port = self._update_port( [ 730.937876] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.937876] env[62952]: _ensure_no_port_binding_failure(port) [ 730.937876] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.937876] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 730.938826] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 9b472956-916e-4f25-a11f-fee191c954c1, please check neutron logs for more information. [ 730.938826] env[62952]: Removing descriptor: 19 [ 731.089375] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.025s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.089998] env[62952]: ERROR nova.compute.manager [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4c5b2393-f9f2-4e98-8c3e-9e834192ee0f, please check neutron logs for more information. [ 731.089998] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Traceback (most recent call last): [ 731.089998] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 731.089998] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] self.driver.spawn(context, instance, image_meta, [ 731.089998] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 731.089998] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 731.089998] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 731.089998] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] vm_ref = self.build_virtual_machine(instance, [ 731.089998] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 731.089998] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] vif_infos = vmwarevif.get_vif_info(self._session, [ 731.089998] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 731.090395] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] for vif in network_info: [ 731.090395] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 731.090395] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] return self._sync_wrapper(fn, *args, **kwargs) [ 731.090395] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 731.090395] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] self.wait() [ 731.090395] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 731.090395] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] self[:] = self._gt.wait() [ 731.090395] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 731.090395] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] return self._exit_event.wait() [ 731.090395] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 731.090395] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] result = hub.switch() [ 731.090395] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 731.090395] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] return self.greenlet.switch() [ 731.090751] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.090751] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] result = function(*args, **kwargs) [ 731.090751] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 731.090751] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] return func(*args, **kwargs) [ 731.090751] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 731.090751] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] raise e [ 731.090751] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.090751] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] nwinfo = self.network_api.allocate_for_instance( [ 731.090751] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 731.090751] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] created_port_ids = self._update_ports_for_instance( [ 731.090751] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 731.090751] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] with excutils.save_and_reraise_exception(): [ 731.090751] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.091113] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] self.force_reraise() [ 731.091113] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.091113] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] raise self.value [ 731.091113] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 731.091113] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] updated_port = self._update_port( [ 731.091113] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.091113] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] _ensure_no_port_binding_failure(port) [ 731.091113] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.091113] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] raise exception.PortBindingFailed(port_id=port['id']) [ 731.091113] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] nova.exception.PortBindingFailed: Binding failed for port 4c5b2393-f9f2-4e98-8c3e-9e834192ee0f, please check neutron logs for more information. [ 731.091113] env[62952]: ERROR nova.compute.manager [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] [ 731.091422] env[62952]: DEBUG nova.compute.utils [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Binding failed for port 4c5b2393-f9f2-4e98-8c3e-9e834192ee0f, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 731.092830] env[62952]: DEBUG nova.compute.manager [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 731.094906] env[62952]: DEBUG nova.compute.manager [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Build of instance 11139a80-26b2-4bc6-af91-18cd3adbe8fe was re-scheduled: Binding failed for port 4c5b2393-f9f2-4e98-8c3e-9e834192ee0f, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 731.095509] env[62952]: DEBUG nova.compute.manager [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 731.095741] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Acquiring lock "refresh_cache-11139a80-26b2-4bc6-af91-18cd3adbe8fe" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.095888] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Acquired lock "refresh_cache-11139a80-26b2-4bc6-af91-18cd3adbe8fe" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.096059] env[62952]: DEBUG nova.network.neutron [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 731.097352] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.883s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.097570] env[62952]: DEBUG nova.objects.instance [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62952) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 731.122752] env[62952]: DEBUG nova.virt.hardware [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 731.123085] env[62952]: DEBUG nova.virt.hardware [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 731.123162] env[62952]: DEBUG nova.virt.hardware [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 731.123429] env[62952]: DEBUG nova.virt.hardware [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 731.123526] env[62952]: DEBUG nova.virt.hardware [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 731.123745] env[62952]: DEBUG nova.virt.hardware [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 731.124027] env[62952]: DEBUG nova.virt.hardware [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 731.124296] env[62952]: DEBUG nova.virt.hardware [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 731.124548] env[62952]: DEBUG nova.virt.hardware [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 731.124909] env[62952]: DEBUG nova.virt.hardware [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 731.125015] env[62952]: DEBUG nova.virt.hardware [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 731.126018] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ee8ceca-f2ac-4884-b41f-48d9f591b875 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.135810] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac6cb21-d49e-4dd5-bef4-43e6fbdaabcb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.150286] env[62952]: ERROR nova.compute.manager [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9b472956-916e-4f25-a11f-fee191c954c1, please check neutron logs for more information. [ 731.150286] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Traceback (most recent call last): [ 731.150286] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 731.150286] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] yield resources [ 731.150286] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 731.150286] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] self.driver.spawn(context, instance, image_meta, [ 731.150286] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 731.150286] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 731.150286] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 731.150286] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] vm_ref = self.build_virtual_machine(instance, [ 731.150286] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 731.150781] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] vif_infos = vmwarevif.get_vif_info(self._session, [ 731.150781] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 731.150781] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] for vif in network_info: [ 731.150781] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 731.150781] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] return self._sync_wrapper(fn, *args, **kwargs) [ 731.150781] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 731.150781] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] self.wait() [ 731.150781] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 731.150781] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] self[:] = self._gt.wait() [ 731.150781] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 731.150781] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] return self._exit_event.wait() [ 731.150781] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 731.150781] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] current.throw(*self._exc) [ 731.151178] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.151178] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] result = function(*args, **kwargs) [ 731.151178] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 731.151178] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] return func(*args, **kwargs) [ 731.151178] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 731.151178] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] raise e [ 731.151178] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.151178] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] nwinfo = self.network_api.allocate_for_instance( [ 731.151178] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 731.151178] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] created_port_ids = self._update_ports_for_instance( [ 731.151178] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 731.151178] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] with excutils.save_and_reraise_exception(): [ 731.151178] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.151643] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] self.force_reraise() [ 731.151643] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.151643] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] raise self.value [ 731.151643] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 731.151643] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] updated_port = self._update_port( [ 731.151643] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.151643] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] _ensure_no_port_binding_failure(port) [ 731.151643] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.151643] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] raise exception.PortBindingFailed(port_id=port['id']) [ 731.151643] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] nova.exception.PortBindingFailed: Binding failed for port 9b472956-916e-4f25-a11f-fee191c954c1, please check neutron logs for more information. [ 731.151643] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] [ 731.151643] env[62952]: INFO nova.compute.manager [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Terminating instance [ 731.152638] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "refresh_cache-f9ed97a4-ec32-415b-958e-d362b622b1d4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.246550] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.382120] env[62952]: DEBUG nova.network.neutron [req-082ac7aa-1ff8-4299-883d-2caec9b8f86f req-ce5e6680-2e80-4b06-9b3e-360cf0147ee5 service nova] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 731.463859] env[62952]: DEBUG nova.network.neutron [req-082ac7aa-1ff8-4299-883d-2caec9b8f86f req-ce5e6680-2e80-4b06-9b3e-360cf0147ee5 service nova] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.618419] env[62952]: DEBUG nova.network.neutron [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 731.662162] env[62952]: DEBUG nova.network.neutron [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.966959] env[62952]: DEBUG oslo_concurrency.lockutils [req-082ac7aa-1ff8-4299-883d-2caec9b8f86f req-ce5e6680-2e80-4b06-9b3e-360cf0147ee5 service nova] Releasing lock "refresh_cache-f9ed97a4-ec32-415b-958e-d362b622b1d4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.967441] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "refresh_cache-f9ed97a4-ec32-415b-958e-d362b622b1d4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.967628] env[62952]: DEBUG nova.network.neutron [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 732.048430] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "5f1eca6f-4cc4-483e-b72a-b479378277fe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.048430] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "5f1eca6f-4cc4-483e-b72a-b479378277fe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.107279] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1677f2eb-80a2-4723-9e93-fce0cbb856c2 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.108426] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.005s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.165126] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Releasing lock "refresh_cache-11139a80-26b2-4bc6-af91-18cd3adbe8fe" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.166423] env[62952]: DEBUG nova.compute.manager [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 732.167032] env[62952]: DEBUG nova.compute.manager [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 732.167032] env[62952]: DEBUG nova.network.neutron [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 732.181984] env[62952]: DEBUG nova.network.neutron [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 732.485093] env[62952]: DEBUG nova.network.neutron [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 732.564921] env[62952]: DEBUG nova.network.neutron [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.684166] env[62952]: DEBUG nova.network.neutron [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.883116] env[62952]: DEBUG nova.compute.manager [req-9902f0e9-47c3-4884-954a-7d46bb856533 req-ac84f5f0-bab9-4522-916e-d5a7aa9fb3a9 service nova] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Received event network-vif-deleted-9b472956-916e-4f25-a11f-fee191c954c1 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 732.936497] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ebbc2e-346d-4949-b610-c73ac99385ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.944610] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beae68f4-78d7-4311-834d-48ebb05bc766 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.976016] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb64be4-61b5-48ab-8a6e-1a96b228a1dc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.983305] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a0d03c-ac82-4022-9c28-4c488b874923 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.996356] env[62952]: DEBUG nova.compute.provider_tree [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.067939] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "refresh_cache-f9ed97a4-ec32-415b-958e-d362b622b1d4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.068430] env[62952]: DEBUG nova.compute.manager [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 733.068624] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 733.068914] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8d344e38-871e-4b8d-a09a-e55bc2aaecb8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.078411] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f82700-57e0-4de4-97ac-2c49dc349599 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.098959] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f9ed97a4-ec32-415b-958e-d362b622b1d4 could not be found. [ 733.099269] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 733.099429] env[62952]: INFO nova.compute.manager [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 733.099700] env[62952]: DEBUG oslo.service.loopingcall [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.099919] env[62952]: DEBUG nova.compute.manager [-] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.100096] env[62952]: DEBUG nova.network.neutron [-] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 733.115413] env[62952]: DEBUG nova.network.neutron [-] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.186938] env[62952]: INFO nova.compute.manager [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] [instance: 11139a80-26b2-4bc6-af91-18cd3adbe8fe] Took 1.02 seconds to deallocate network for instance. [ 733.500068] env[62952]: DEBUG nova.scheduler.client.report [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 733.617691] env[62952]: DEBUG nova.network.neutron [-] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.006030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.897s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.007883] env[62952]: ERROR nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e860ee81-765c-4885-979c-9d8c922d8a80, please check neutron logs for more information. [ 734.007883] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Traceback (most recent call last): [ 734.007883] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 734.007883] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] self.driver.spawn(context, instance, image_meta, [ 734.007883] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 734.007883] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 734.007883] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 734.007883] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] vm_ref = self.build_virtual_machine(instance, [ 734.007883] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 734.007883] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] vif_infos = vmwarevif.get_vif_info(self._session, [ 734.007883] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 734.008285] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] for vif in network_info: [ 734.008285] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 734.008285] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] return self._sync_wrapper(fn, *args, **kwargs) [ 734.008285] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 734.008285] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] self.wait() [ 734.008285] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 734.008285] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] self[:] = self._gt.wait() [ 734.008285] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 734.008285] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] return self._exit_event.wait() [ 734.008285] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 734.008285] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] result = hub.switch() [ 734.008285] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 734.008285] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] return self.greenlet.switch() [ 734.008590] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.008590] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] result = function(*args, **kwargs) [ 734.008590] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 734.008590] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] return func(*args, **kwargs) [ 734.008590] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.008590] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] raise e [ 734.008590] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.008590] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] nwinfo = self.network_api.allocate_for_instance( [ 734.008590] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 734.008590] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] created_port_ids = self._update_ports_for_instance( [ 734.008590] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 734.008590] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] with excutils.save_and_reraise_exception(): [ 734.008590] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.008896] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] self.force_reraise() [ 734.008896] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.008896] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] raise self.value [ 734.008896] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 734.008896] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] updated_port = self._update_port( [ 734.008896] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.008896] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] _ensure_no_port_binding_failure(port) [ 734.008896] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.008896] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] raise exception.PortBindingFailed(port_id=port['id']) [ 734.008896] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] nova.exception.PortBindingFailed: Binding failed for port e860ee81-765c-4885-979c-9d8c922d8a80, please check neutron logs for more information. [ 734.008896] env[62952]: ERROR nova.compute.manager [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] [ 734.009532] env[62952]: DEBUG nova.compute.utils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Binding failed for port e860ee81-765c-4885-979c-9d8c922d8a80, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 734.009532] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.571s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.009532] env[62952]: DEBUG nova.objects.instance [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Lazy-loading 'resources' on Instance uuid f075161d-7447-4601-a06c-514fc4dd2eb9 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 734.010245] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Build of instance 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b was re-scheduled: Binding failed for port e860ee81-765c-4885-979c-9d8c922d8a80, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 734.010655] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 734.011190] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquiring lock "refresh_cache-7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.011190] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquired lock "refresh_cache-7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.011190] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 734.121357] env[62952]: INFO nova.compute.manager [-] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Took 1.02 seconds to deallocate network for instance. [ 734.122934] env[62952]: DEBUG nova.compute.claims [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 734.123121] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.218376] env[62952]: INFO nova.scheduler.client.report [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Deleted allocations for instance 11139a80-26b2-4bc6-af91-18cd3adbe8fe [ 734.536207] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 734.618414] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.728142] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ffc410e-f374-4f3d-8bec-f86ab467c475 tempest-DeleteServersAdminTestJSON-1584051569 tempest-DeleteServersAdminTestJSON-1584051569-project-member] Lock "11139a80-26b2-4bc6-af91-18cd3adbe8fe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.847s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.833886] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981eee04-5e7b-4a4a-84a3-87f269f190c0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.843039] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8adc94d3-0400-49e9-bd62-3f1a0fa8a129 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.873376] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a13ccdf3-1003-461a-bce1-7d4200a923a3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.880498] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-100f3503-6613-4497-b078-4774bc9eee04 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.893236] env[62952]: DEBUG nova.compute.provider_tree [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.123485] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Releasing lock "refresh_cache-7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.123865] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 735.123933] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 735.124073] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 735.139904] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 735.231368] env[62952]: DEBUG nova.compute.manager [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 735.396498] env[62952]: DEBUG nova.scheduler.client.report [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 735.643154] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.753870] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.902920] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.894s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.905433] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.368s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.924046] env[62952]: INFO nova.scheduler.client.report [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Deleted allocations for instance f075161d-7447-4601-a06c-514fc4dd2eb9 [ 736.145804] env[62952]: INFO nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b] Took 1.02 seconds to deallocate network for instance. [ 736.435511] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58b91ea8-01b3-4fce-8deb-61f0506bef84 tempest-ServerShowV254Test-68451117 tempest-ServerShowV254Test-68451117-project-member] Lock "f075161d-7447-4601-a06c-514fc4dd2eb9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.672s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.732807] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb8d9629-c844-4cb8-a4da-511306c7a012 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.741326] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a111057-17b8-4409-a288-84b0af49d73d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.772718] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d48b57-1e53-4252-994c-550071d510ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.780229] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b19923b-2596-44e1-bc1d-d197de96ac1c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.796601] env[62952]: DEBUG nova.compute.provider_tree [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.188815] env[62952]: INFO nova.scheduler.client.report [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Deleted allocations for instance 7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b [ 737.300129] env[62952]: DEBUG nova.scheduler.client.report [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.700267] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "7bdbcd4e-9ebd-43e2-91cb-dd365a605b7b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.863s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.805706] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.900s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.808122] env[62952]: ERROR nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e5af67a4-3935-4376-9260-4fd7954e2d9b, please check neutron logs for more information. [ 737.808122] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Traceback (most recent call last): [ 737.808122] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 737.808122] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] self.driver.spawn(context, instance, image_meta, [ 737.808122] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 737.808122] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 737.808122] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 737.808122] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] vm_ref = self.build_virtual_machine(instance, [ 737.808122] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 737.808122] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] vif_infos = vmwarevif.get_vif_info(self._session, [ 737.808122] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 737.808403] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] for vif in network_info: [ 737.808403] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 737.808403] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] return self._sync_wrapper(fn, *args, **kwargs) [ 737.808403] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 737.808403] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] self.wait() [ 737.808403] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 737.808403] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] self[:] = self._gt.wait() [ 737.808403] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 737.808403] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] return self._exit_event.wait() [ 737.808403] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 737.808403] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] current.throw(*self._exc) [ 737.808403] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.808403] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] result = function(*args, **kwargs) [ 737.808714] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 737.808714] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] return func(*args, **kwargs) [ 737.808714] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.808714] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] raise e [ 737.808714] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.808714] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] nwinfo = self.network_api.allocate_for_instance( [ 737.808714] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 737.808714] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] created_port_ids = self._update_ports_for_instance( [ 737.808714] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 737.808714] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] with excutils.save_and_reraise_exception(): [ 737.808714] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.808714] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] self.force_reraise() [ 737.808714] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.809271] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] raise self.value [ 737.809271] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 737.809271] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] updated_port = self._update_port( [ 737.809271] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.809271] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] _ensure_no_port_binding_failure(port) [ 737.809271] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.809271] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] raise exception.PortBindingFailed(port_id=port['id']) [ 737.809271] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] nova.exception.PortBindingFailed: Binding failed for port e5af67a4-3935-4376-9260-4fd7954e2d9b, please check neutron logs for more information. [ 737.809271] env[62952]: ERROR nova.compute.manager [instance: 36f778a3-d795-467c-92d9-97e07b04956c] [ 737.809271] env[62952]: DEBUG nova.compute.utils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Binding failed for port e5af67a4-3935-4376-9260-4fd7954e2d9b, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 737.814572] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.553s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.817682] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Build of instance 36f778a3-d795-467c-92d9-97e07b04956c was re-scheduled: Binding failed for port e5af67a4-3935-4376-9260-4fd7954e2d9b, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 737.817682] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 737.817682] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquiring lock "refresh_cache-36f778a3-d795-467c-92d9-97e07b04956c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.818019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquired lock "refresh_cache-36f778a3-d795-467c-92d9-97e07b04956c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.818284] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 738.204817] env[62952]: DEBUG nova.compute.manager [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 738.359534] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 738.506706] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.700491] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c48546-9b1c-49a4-83da-164f30dac7a1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.710304] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bcc4f9a-1582-48db-b889-faa8bae32aed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.752082] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f613ee1-2281-4511-b696-41274309ef55 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.760518] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbf52d11-ad1b-46ce-8524-2ac971babc28 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.766924] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.776990] env[62952]: DEBUG nova.compute.provider_tree [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.008648] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Releasing lock "refresh_cache-36f778a3-d795-467c-92d9-97e07b04956c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.008927] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 739.009136] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.009311] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 739.025405] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 739.279949] env[62952]: DEBUG nova.scheduler.client.report [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.528058] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.785531] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.974s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.786174] env[62952]: ERROR nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7b9602b0-50f1-4711-826f-2fa2c7dea325, please check neutron logs for more information. [ 739.786174] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Traceback (most recent call last): [ 739.786174] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 739.786174] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] self.driver.spawn(context, instance, image_meta, [ 739.786174] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 739.786174] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.786174] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.786174] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] vm_ref = self.build_virtual_machine(instance, [ 739.786174] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.786174] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.786174] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.786467] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] for vif in network_info: [ 739.786467] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.786467] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] return self._sync_wrapper(fn, *args, **kwargs) [ 739.786467] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.786467] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] self.wait() [ 739.786467] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.786467] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] self[:] = self._gt.wait() [ 739.786467] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.786467] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] return self._exit_event.wait() [ 739.786467] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 739.786467] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] current.throw(*self._exc) [ 739.786467] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.786467] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] result = function(*args, **kwargs) [ 739.786773] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 739.786773] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] return func(*args, **kwargs) [ 739.786773] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.786773] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] raise e [ 739.786773] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.786773] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] nwinfo = self.network_api.allocate_for_instance( [ 739.786773] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 739.786773] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] created_port_ids = self._update_ports_for_instance( [ 739.786773] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 739.786773] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] with excutils.save_and_reraise_exception(): [ 739.786773] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.786773] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] self.force_reraise() [ 739.786773] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.787085] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] raise self.value [ 739.787085] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 739.787085] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] updated_port = self._update_port( [ 739.787085] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.787085] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] _ensure_no_port_binding_failure(port) [ 739.787085] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.787085] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] raise exception.PortBindingFailed(port_id=port['id']) [ 739.787085] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] nova.exception.PortBindingFailed: Binding failed for port 7b9602b0-50f1-4711-826f-2fa2c7dea325, please check neutron logs for more information. [ 739.787085] env[62952]: ERROR nova.compute.manager [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] [ 739.787085] env[62952]: DEBUG nova.compute.utils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Binding failed for port 7b9602b0-50f1-4711-826f-2fa2c7dea325, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 739.789075] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.792911] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Build of instance 7419de8d-5198-4152-a410-818af9c3b8e2 was re-scheduled: Binding failed for port 7b9602b0-50f1-4711-826f-2fa2c7dea325, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 739.796033] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 739.796033] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquiring lock "refresh_cache-7419de8d-5198-4152-a410-818af9c3b8e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.796033] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Acquired lock "refresh_cache-7419de8d-5198-4152-a410-818af9c3b8e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.796033] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 740.031105] env[62952]: INFO nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 36f778a3-d795-467c-92d9-97e07b04956c] Took 1.02 seconds to deallocate network for instance. [ 740.325563] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 740.483664] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.700365] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc5e6d44-8f4d-4eb5-bdd8-8a4f0070c44b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.708010] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f755ab-48b0-4329-a4db-6064dbaf953a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.742340] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee767c3a-da47-4d3c-b285-7644e5d6149e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.752837] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8822ec55-0796-4670-a24d-98715a05c73e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.768026] env[62952]: DEBUG nova.compute.provider_tree [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.987474] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Releasing lock "refresh_cache-7419de8d-5198-4152-a410-818af9c3b8e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.987474] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 740.987474] env[62952]: DEBUG nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.987474] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 741.011301] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.071133] env[62952]: INFO nova.scheduler.client.report [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Deleted allocations for instance 36f778a3-d795-467c-92d9-97e07b04956c [ 741.271015] env[62952]: DEBUG nova.scheduler.client.report [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.518445] env[62952]: DEBUG nova.network.neutron [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.581762] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "36f778a3-d795-467c-92d9-97e07b04956c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.693s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.785475] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.995s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.785475] env[62952]: ERROR nova.compute.manager [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 141f9cd0-39d5-42f1-944e-bded6127f741, please check neutron logs for more information. [ 741.785475] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Traceback (most recent call last): [ 741.785475] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 741.785475] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] self.driver.spawn(context, instance, image_meta, [ 741.785475] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 741.785475] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] self._vmops.spawn(context, instance, image_meta, injected_files, [ 741.785475] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 741.785475] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] vm_ref = self.build_virtual_machine(instance, [ 741.786155] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 741.786155] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] vif_infos = vmwarevif.get_vif_info(self._session, [ 741.786155] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 741.786155] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] for vif in network_info: [ 741.786155] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 741.786155] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] return self._sync_wrapper(fn, *args, **kwargs) [ 741.786155] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 741.786155] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] self.wait() [ 741.786155] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 741.786155] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] self[:] = self._gt.wait() [ 741.786155] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 741.786155] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] return self._exit_event.wait() [ 741.786155] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 741.787219] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] current.throw(*self._exc) [ 741.787219] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.787219] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] result = function(*args, **kwargs) [ 741.787219] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 741.787219] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] return func(*args, **kwargs) [ 741.787219] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.787219] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] raise e [ 741.787219] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.787219] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] nwinfo = self.network_api.allocate_for_instance( [ 741.787219] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 741.787219] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] created_port_ids = self._update_ports_for_instance( [ 741.787219] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 741.787219] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] with excutils.save_and_reraise_exception(): [ 741.787859] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.787859] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] self.force_reraise() [ 741.787859] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.787859] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] raise self.value [ 741.787859] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 741.787859] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] updated_port = self._update_port( [ 741.787859] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.787859] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] _ensure_no_port_binding_failure(port) [ 741.787859] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.787859] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] raise exception.PortBindingFailed(port_id=port['id']) [ 741.787859] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] nova.exception.PortBindingFailed: Binding failed for port 141f9cd0-39d5-42f1-944e-bded6127f741, please check neutron logs for more information. [ 741.787859] env[62952]: ERROR nova.compute.manager [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] [ 741.788386] env[62952]: DEBUG nova.compute.utils [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Binding failed for port 141f9cd0-39d5-42f1-944e-bded6127f741, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 741.788386] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 19.675s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.792021] env[62952]: DEBUG nova.compute.manager [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Build of instance 0cdbad9a-9d52-45b0-8267-e181a5d38134 was re-scheduled: Binding failed for port 141f9cd0-39d5-42f1-944e-bded6127f741, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 741.792021] env[62952]: DEBUG nova.compute.manager [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 741.792021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Acquiring lock "refresh_cache-0cdbad9a-9d52-45b0-8267-e181a5d38134" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.792021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Acquired lock "refresh_cache-0cdbad9a-9d52-45b0-8267-e181a5d38134" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.792224] env[62952]: DEBUG nova.network.neutron [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 742.025155] env[62952]: INFO nova.compute.manager [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] [instance: 7419de8d-5198-4152-a410-818af9c3b8e2] Took 1.04 seconds to deallocate network for instance. [ 742.084836] env[62952]: DEBUG nova.compute.manager [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 742.310995] env[62952]: DEBUG nova.network.neutron [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 742.391519] env[62952]: DEBUG nova.network.neutron [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.603372] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.818581] env[62952]: WARNING nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 80a30ea8-e260-4b92-9099-be374333896f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 742.891805] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Releasing lock "refresh_cache-0cdbad9a-9d52-45b0-8267-e181a5d38134" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.892090] env[62952]: DEBUG nova.compute.manager [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 742.892286] env[62952]: DEBUG nova.compute.manager [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 742.892458] env[62952]: DEBUG nova.network.neutron [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 742.916810] env[62952]: DEBUG nova.network.neutron [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 743.028356] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.028606] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.063263] env[62952]: INFO nova.scheduler.client.report [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Deleted allocations for instance 7419de8d-5198-4152-a410-818af9c3b8e2 [ 743.300088] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Acquiring lock "219cfb14-ca73-45d3-86e8-ed4642d7d480" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.300088] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Lock "219cfb14-ca73-45d3-86e8-ed4642d7d480" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.329043] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 7419de8d-5198-4152-a410-818af9c3b8e2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 743.420492] env[62952]: DEBUG nova.network.neutron [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.574913] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c4e671ee-c7fd-4e62-a0d3-312e8facf0ba tempest-ListServersNegativeTestJSON-839764434 tempest-ListServersNegativeTestJSON-839764434-project-member] Lock "7419de8d-5198-4152-a410-818af9c3b8e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.625s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.833503] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 0cdbad9a-9d52-45b0-8267-e181a5d38134 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 743.833787] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 4bc891c1-e7e1-4196-9770-9bb42ed7499c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 743.833787] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance f9ed97a4-ec32-415b-958e-d362b622b1d4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 743.833900] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 268e5249-0156-45dd-bbbd-10105d16f269 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 743.923132] env[62952]: INFO nova.compute.manager [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] [instance: 0cdbad9a-9d52-45b0-8267-e181a5d38134] Took 1.03 seconds to deallocate network for instance. [ 744.077511] env[62952]: DEBUG nova.compute.manager [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 744.338982] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 5922d142-0b6f-4479-abcb-0c067a1ef837 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 744.601288] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.841724] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance a458a658-96f3-41b2-a9ee-2a6dd52bfeb4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 744.956018] env[62952]: INFO nova.scheduler.client.report [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Deleted allocations for instance 0cdbad9a-9d52-45b0-8267-e181a5d38134 [ 745.345012] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance dd88d70d-17a3-4059-9092-5d6cec8b1ae4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 745.461218] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e8905b65-f29c-4efa-b04e-4ba8f35848bb tempest-InstanceActionsNegativeTestJSON-1282841801 tempest-InstanceActionsNegativeTestJSON-1282841801-project-member] Lock "0cdbad9a-9d52-45b0-8267-e181a5d38134" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.359s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.850285] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 472e60f1-9b6c-415e-8a92-19235ad2e547 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 745.965299] env[62952]: DEBUG nova.compute.manager [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 746.353890] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 746.491446] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.860349] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance da96503a-8ae8-4b68-a782-631ee98754f7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 747.365706] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 7f77fa00-27be-4a75-aac0-e1918f817837 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 747.828028] env[62952]: DEBUG oslo_concurrency.lockutils [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquiring lock "bc369b94-2f40-444e-88d7-069fd33e2646" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.828131] env[62952]: DEBUG oslo_concurrency.lockutils [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Lock "bc369b94-2f40-444e-88d7-069fd33e2646" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.869168] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 879dc564-0a5e-48aa-92ff-b02435445a68 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 748.372025] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 3822c92d-560b-4d61-92e9-7c694bc7a33a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 748.878613] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 9979dd1d-9e99-47cf-a81a-fd7d89531f51 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 749.383378] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance ebff0c87-8e58-4a76-a350-43a3d660debd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 749.886466] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 1ed9de84-6178-4173-84ad-dc115be7a1fa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 749.948383] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.948383] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.162909] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "58e703ea-7ccf-4114-a5a2-49533390ed37" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.163160] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "58e703ea-7ccf-4114-a5a2-49533390ed37" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.389555] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 1d6447ab-dc07-459f-9124-d881e0da7cf1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 750.768238] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "f906dc47-12cb-46f5-ae5c-53cfcc8765a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.768436] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "f906dc47-12cb-46f5-ae5c-53cfcc8765a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.892963] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 9c5167eb-3a15-47ff-923e-c83cc3529cc8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 751.396318] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 9be8f7f5-80eb-4e38-9501-e76abd7227db has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 751.899991] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 8ac98d96-139a-4a03-94f3-178fafa88503 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 752.403647] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance f4f8bee4-bf92-4321-afdc-16f7fb17dbae has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 752.906912] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 5f1eca6f-4cc4-483e-b72a-b479378277fe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 753.409561] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 9cc7bc2f-ff22-43cf-be79-ff8c6426115b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 753.409821] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 753.409975] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 753.708405] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8717a152-54fa-4374-9517-34f320ac42cb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.715909] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e852f0-b51b-4f13-9be8-0292f4507abd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.745407] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e81f5c-aadb-467c-8c59-bb520fb1eff3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.752425] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb5770f8-adb8-489b-a9b3-a969fdb87c48 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.765034] env[62952]: DEBUG nova.compute.provider_tree [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.268976] env[62952]: DEBUG nova.scheduler.client.report [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.773555] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 754.773810] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 12.988s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.774104] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.569s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.774288] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.776266] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.583s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.777806] env[62952]: INFO nova.compute.claims [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 754.796570] env[62952]: INFO nova.scheduler.client.report [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Deleted allocations for instance 80a30ea8-e260-4b92-9099-be374333896f [ 755.303863] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e80296f7-7f81-4443-91d0-97aa6da105f4 tempest-ServerDiagnosticsV248Test-916910100 tempest-ServerDiagnosticsV248Test-916910100-project-member] Lock "80a30ea8-e260-4b92-9099-be374333896f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.840s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.109685] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3b392c5-bcc1-4064-a21f-9cddc668a358 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.117318] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17dbcb69-43d7-452a-bef9-6c2304532fec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.146441] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c68e0c-3070-49e2-a2fc-1c6a4415c790 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.154083] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a232c0-290d-48b7-9149-72bd23bde5e3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.167593] env[62952]: DEBUG nova.compute.provider_tree [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.671806] env[62952]: DEBUG nova.scheduler.client.report [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 757.176284] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.400s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.176805] env[62952]: DEBUG nova.compute.manager [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 757.179664] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.141s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.685040] env[62952]: DEBUG nova.compute.utils [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 757.695940] env[62952]: DEBUG nova.compute.manager [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 757.698683] env[62952]: DEBUG nova.compute.manager [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 757.698900] env[62952]: DEBUG nova.network.neutron [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 757.756751] env[62952]: DEBUG nova.policy [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '71f753644029446085fc97e284efb94a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e45ef0f648294454b86d27b13af6a4cf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 758.045609] env[62952]: DEBUG nova.network.neutron [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Successfully created port: 00a9284f-3e43-4013-aa09-d38cfc6414ce {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 758.101501] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98a3bd49-dcc3-4eed-81d7-faf5fb8e4caf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.109561] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed27732f-85eb-4500-bbda-5448a4d74717 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.141342] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b81481-9604-42f5-95ef-001db6030983 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.150258] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-107affd4-c7cc-47d0-b5c3-5533c02d1185 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.164364] env[62952]: DEBUG nova.compute.provider_tree [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.668092] env[62952]: DEBUG nova.scheduler.client.report [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 758.708881] env[62952]: DEBUG nova.compute.manager [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 758.742174] env[62952]: DEBUG nova.compute.manager [req-a41e76dd-abe8-4d1b-a6ba-00b9742d9fde req-9f9377f8-7c05-4605-94e0-30c284588ffe service nova] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Received event network-changed-00a9284f-3e43-4013-aa09-d38cfc6414ce {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 758.742289] env[62952]: DEBUG nova.compute.manager [req-a41e76dd-abe8-4d1b-a6ba-00b9742d9fde req-9f9377f8-7c05-4605-94e0-30c284588ffe service nova] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Refreshing instance network info cache due to event network-changed-00a9284f-3e43-4013-aa09-d38cfc6414ce. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 758.742883] env[62952]: DEBUG oslo_concurrency.lockutils [req-a41e76dd-abe8-4d1b-a6ba-00b9742d9fde req-9f9377f8-7c05-4605-94e0-30c284588ffe service nova] Acquiring lock "refresh_cache-5922d142-0b6f-4479-abcb-0c067a1ef837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.742883] env[62952]: DEBUG oslo_concurrency.lockutils [req-a41e76dd-abe8-4d1b-a6ba-00b9742d9fde req-9f9377f8-7c05-4605-94e0-30c284588ffe service nova] Acquired lock "refresh_cache-5922d142-0b6f-4479-abcb-0c067a1ef837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.742883] env[62952]: DEBUG nova.network.neutron [req-a41e76dd-abe8-4d1b-a6ba-00b9742d9fde req-9f9377f8-7c05-4605-94e0-30c284588ffe service nova] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Refreshing network info cache for port 00a9284f-3e43-4013-aa09-d38cfc6414ce {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 758.746457] env[62952]: DEBUG nova.virt.hardware [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 758.746672] env[62952]: DEBUG nova.virt.hardware [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 758.746832] env[62952]: DEBUG nova.virt.hardware [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 758.749070] env[62952]: DEBUG nova.virt.hardware [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 758.749070] env[62952]: DEBUG nova.virt.hardware [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 758.749070] env[62952]: DEBUG nova.virt.hardware [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 758.749070] env[62952]: DEBUG nova.virt.hardware [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 758.749070] env[62952]: DEBUG nova.virt.hardware [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 758.750725] env[62952]: DEBUG nova.virt.hardware [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 758.750725] env[62952]: DEBUG nova.virt.hardware [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 758.750725] env[62952]: DEBUG nova.virt.hardware [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 758.750725] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1702c8e0-d02f-4cf1-9139-de52cd45a4f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.760973] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b2296a6-bf4f-443f-afcb-e7de34164714 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.985990] env[62952]: ERROR nova.compute.manager [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 00a9284f-3e43-4013-aa09-d38cfc6414ce, please check neutron logs for more information. [ 758.985990] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 758.985990] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.985990] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 758.985990] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 758.985990] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 758.985990] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 758.985990] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 758.985990] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.985990] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 758.985990] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.985990] env[62952]: ERROR nova.compute.manager raise self.value [ 758.985990] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 758.985990] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 758.985990] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.985990] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 758.986475] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.986475] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 758.986475] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 00a9284f-3e43-4013-aa09-d38cfc6414ce, please check neutron logs for more information. [ 758.986475] env[62952]: ERROR nova.compute.manager [ 758.986475] env[62952]: Traceback (most recent call last): [ 758.986475] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 758.986475] env[62952]: listener.cb(fileno) [ 758.986475] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.986475] env[62952]: result = function(*args, **kwargs) [ 758.986475] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 758.986475] env[62952]: return func(*args, **kwargs) [ 758.986475] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.986475] env[62952]: raise e [ 758.986475] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.986475] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 758.986475] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 758.986475] env[62952]: created_port_ids = self._update_ports_for_instance( [ 758.986475] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 758.986475] env[62952]: with excutils.save_and_reraise_exception(): [ 758.986475] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.986475] env[62952]: self.force_reraise() [ 758.986475] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.986475] env[62952]: raise self.value [ 758.986475] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 758.986475] env[62952]: updated_port = self._update_port( [ 758.986475] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.986475] env[62952]: _ensure_no_port_binding_failure(port) [ 758.986475] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.986475] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 758.987310] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 00a9284f-3e43-4013-aa09-d38cfc6414ce, please check neutron logs for more information. [ 758.987310] env[62952]: Removing descriptor: 19 [ 758.987310] env[62952]: ERROR nova.compute.manager [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 00a9284f-3e43-4013-aa09-d38cfc6414ce, please check neutron logs for more information. [ 758.987310] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Traceback (most recent call last): [ 758.987310] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 758.987310] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] yield resources [ 758.987310] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 758.987310] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] self.driver.spawn(context, instance, image_meta, [ 758.987310] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 758.987310] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] self._vmops.spawn(context, instance, image_meta, injected_files, [ 758.987310] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 758.987310] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] vm_ref = self.build_virtual_machine(instance, [ 758.987703] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 758.987703] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] vif_infos = vmwarevif.get_vif_info(self._session, [ 758.987703] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 758.987703] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] for vif in network_info: [ 758.987703] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 758.987703] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] return self._sync_wrapper(fn, *args, **kwargs) [ 758.987703] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 758.987703] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] self.wait() [ 758.987703] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 758.987703] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] self[:] = self._gt.wait() [ 758.987703] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 758.987703] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] return self._exit_event.wait() [ 758.987703] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 758.988054] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] result = hub.switch() [ 758.988054] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 758.988054] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] return self.greenlet.switch() [ 758.988054] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.988054] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] result = function(*args, **kwargs) [ 758.988054] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 758.988054] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] return func(*args, **kwargs) [ 758.988054] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.988054] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] raise e [ 758.988054] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.988054] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] nwinfo = self.network_api.allocate_for_instance( [ 758.988054] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 758.988054] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] created_port_ids = self._update_ports_for_instance( [ 758.988438] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 758.988438] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] with excutils.save_and_reraise_exception(): [ 758.988438] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.988438] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] self.force_reraise() [ 758.988438] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.988438] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] raise self.value [ 758.988438] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 758.988438] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] updated_port = self._update_port( [ 758.988438] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.988438] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] _ensure_no_port_binding_failure(port) [ 758.988438] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.988438] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] raise exception.PortBindingFailed(port_id=port['id']) [ 758.988822] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] nova.exception.PortBindingFailed: Binding failed for port 00a9284f-3e43-4013-aa09-d38cfc6414ce, please check neutron logs for more information. [ 758.988822] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] [ 758.988822] env[62952]: INFO nova.compute.manager [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Terminating instance [ 758.989428] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Acquiring lock "refresh_cache-5922d142-0b6f-4479-abcb-0c067a1ef837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.173843] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.994s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.174499] env[62952]: ERROR nova.compute.manager [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 52ae2518-4f0c-4959-9d2c-00f38c4749fb, please check neutron logs for more information. [ 759.174499] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Traceback (most recent call last): [ 759.174499] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 759.174499] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] self.driver.spawn(context, instance, image_meta, [ 759.174499] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 759.174499] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 759.174499] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 759.174499] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] vm_ref = self.build_virtual_machine(instance, [ 759.174499] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 759.174499] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] vif_infos = vmwarevif.get_vif_info(self._session, [ 759.174499] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 759.174806] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] for vif in network_info: [ 759.174806] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 759.174806] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] return self._sync_wrapper(fn, *args, **kwargs) [ 759.174806] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 759.174806] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] self.wait() [ 759.174806] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 759.174806] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] self[:] = self._gt.wait() [ 759.174806] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 759.174806] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] return self._exit_event.wait() [ 759.174806] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 759.174806] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] current.throw(*self._exc) [ 759.174806] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.174806] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] result = function(*args, **kwargs) [ 759.175139] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 759.175139] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] return func(*args, **kwargs) [ 759.175139] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 759.175139] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] raise e [ 759.175139] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.175139] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] nwinfo = self.network_api.allocate_for_instance( [ 759.175139] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 759.175139] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] created_port_ids = self._update_ports_for_instance( [ 759.175139] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 759.175139] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] with excutils.save_and_reraise_exception(): [ 759.175139] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.175139] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] self.force_reraise() [ 759.175139] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.175442] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] raise self.value [ 759.175442] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 759.175442] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] updated_port = self._update_port( [ 759.175442] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.175442] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] _ensure_no_port_binding_failure(port) [ 759.175442] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.175442] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] raise exception.PortBindingFailed(port_id=port['id']) [ 759.175442] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] nova.exception.PortBindingFailed: Binding failed for port 52ae2518-4f0c-4959-9d2c-00f38c4749fb, please check neutron logs for more information. [ 759.175442] env[62952]: ERROR nova.compute.manager [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] [ 759.175442] env[62952]: DEBUG nova.compute.utils [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Binding failed for port 52ae2518-4f0c-4959-9d2c-00f38c4749fb, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 759.176905] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.321s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.182248] env[62952]: DEBUG nova.compute.manager [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Build of instance 4bc891c1-e7e1-4196-9770-9bb42ed7499c was re-scheduled: Binding failed for port 52ae2518-4f0c-4959-9d2c-00f38c4749fb, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 759.182708] env[62952]: DEBUG nova.compute.manager [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 759.182930] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Acquiring lock "refresh_cache-4bc891c1-e7e1-4196-9770-9bb42ed7499c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.183092] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Acquired lock "refresh_cache-4bc891c1-e7e1-4196-9770-9bb42ed7499c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.183253] env[62952]: DEBUG nova.network.neutron [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 759.268420] env[62952]: DEBUG nova.network.neutron [req-a41e76dd-abe8-4d1b-a6ba-00b9742d9fde req-9f9377f8-7c05-4605-94e0-30c284588ffe service nova] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.344665] env[62952]: DEBUG nova.network.neutron [req-a41e76dd-abe8-4d1b-a6ba-00b9742d9fde req-9f9377f8-7c05-4605-94e0-30c284588ffe service nova] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.706271] env[62952]: DEBUG nova.network.neutron [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.799469] env[62952]: DEBUG nova.network.neutron [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.847992] env[62952]: DEBUG oslo_concurrency.lockutils [req-a41e76dd-abe8-4d1b-a6ba-00b9742d9fde req-9f9377f8-7c05-4605-94e0-30c284588ffe service nova] Releasing lock "refresh_cache-5922d142-0b6f-4479-abcb-0c067a1ef837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.848136] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Acquired lock "refresh_cache-5922d142-0b6f-4479-abcb-0c067a1ef837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.848319] env[62952]: DEBUG nova.network.neutron [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 760.089806] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c13636-3973-4c4b-908b-feaaee4cf4a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.099573] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c413c4-3d67-4c85-8046-c8e0576840f1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.130282] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2b480c-948f-42f3-873a-9d0669ee6466 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.138383] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452ac108-a7fa-403f-a94b-b7ed79ca53a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.152528] env[62952]: DEBUG nova.compute.provider_tree [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.309413] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Releasing lock "refresh_cache-4bc891c1-e7e1-4196-9770-9bb42ed7499c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.309413] env[62952]: DEBUG nova.compute.manager [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 760.309552] env[62952]: DEBUG nova.compute.manager [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 760.309660] env[62952]: DEBUG nova.network.neutron [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 760.324692] env[62952]: DEBUG nova.network.neutron [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 760.366472] env[62952]: DEBUG nova.network.neutron [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 760.458380] env[62952]: DEBUG nova.network.neutron [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.656009] env[62952]: DEBUG nova.scheduler.client.report [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 760.827153] env[62952]: DEBUG nova.network.neutron [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.961684] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Releasing lock "refresh_cache-5922d142-0b6f-4479-abcb-0c067a1ef837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.962201] env[62952]: DEBUG nova.compute.manager [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 760.962402] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 760.962703] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bfc70052-8be5-438e-8f35-ea80d3c7ad61 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.972276] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-772de884-87c8-4bd2-bd97-47c77a200078 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.995243] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5922d142-0b6f-4479-abcb-0c067a1ef837 could not be found. [ 760.995484] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 760.995718] env[62952]: INFO nova.compute.manager [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Took 0.03 seconds to destroy the instance on the hypervisor. [ 760.996014] env[62952]: DEBUG oslo.service.loopingcall [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 760.996250] env[62952]: DEBUG nova.compute.manager [-] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 760.996342] env[62952]: DEBUG nova.network.neutron [-] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 761.015685] env[62952]: DEBUG nova.network.neutron [-] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 761.078202] env[62952]: DEBUG nova.compute.manager [req-47ed1c78-a245-41d5-bd0b-ec432d0307fa req-15195003-3368-4fd3-bb7b-81c4c5116e43 service nova] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Received event network-vif-deleted-00a9284f-3e43-4013-aa09-d38cfc6414ce {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 761.162746] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.985s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.162746] env[62952]: ERROR nova.compute.manager [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 249319dd-b412-4d3c-9b70-05b017a04e10, please check neutron logs for more information. [ 761.162746] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Traceback (most recent call last): [ 761.162746] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 761.162746] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] self.driver.spawn(context, instance, image_meta, [ 761.162746] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 761.162746] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] self._vmops.spawn(context, instance, image_meta, injected_files, [ 761.162746] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 761.162746] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] vm_ref = self.build_virtual_machine(instance, [ 761.163138] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 761.163138] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] vif_infos = vmwarevif.get_vif_info(self._session, [ 761.163138] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 761.163138] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] for vif in network_info: [ 761.163138] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 761.163138] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] return self._sync_wrapper(fn, *args, **kwargs) [ 761.163138] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 761.163138] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] self.wait() [ 761.163138] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 761.163138] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] self[:] = self._gt.wait() [ 761.163138] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 761.163138] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] return self._exit_event.wait() [ 761.163138] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 761.163462] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] result = hub.switch() [ 761.163462] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 761.163462] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] return self.greenlet.switch() [ 761.163462] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.163462] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] result = function(*args, **kwargs) [ 761.163462] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 761.163462] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] return func(*args, **kwargs) [ 761.163462] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.163462] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] raise e [ 761.163462] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.163462] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] nwinfo = self.network_api.allocate_for_instance( [ 761.163462] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 761.163462] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] created_port_ids = self._update_ports_for_instance( [ 761.163777] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 761.163777] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] with excutils.save_and_reraise_exception(): [ 761.163777] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.163777] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] self.force_reraise() [ 761.163777] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.163777] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] raise self.value [ 761.163777] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 761.163777] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] updated_port = self._update_port( [ 761.163777] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.163777] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] _ensure_no_port_binding_failure(port) [ 761.163777] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.163777] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] raise exception.PortBindingFailed(port_id=port['id']) [ 761.164081] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] nova.exception.PortBindingFailed: Binding failed for port 249319dd-b412-4d3c-9b70-05b017a04e10, please check neutron logs for more information. [ 761.164081] env[62952]: ERROR nova.compute.manager [instance: 268e5249-0156-45dd-bbbd-10105d16f269] [ 761.164081] env[62952]: DEBUG nova.compute.utils [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Binding failed for port 249319dd-b412-4d3c-9b70-05b017a04e10, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 761.164478] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.918s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.166147] env[62952]: INFO nova.compute.claims [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 761.168887] env[62952]: DEBUG nova.compute.manager [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Build of instance 268e5249-0156-45dd-bbbd-10105d16f269 was re-scheduled: Binding failed for port 249319dd-b412-4d3c-9b70-05b017a04e10, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 761.169462] env[62952]: DEBUG nova.compute.manager [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 761.169845] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Acquiring lock "refresh_cache-268e5249-0156-45dd-bbbd-10105d16f269" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.169845] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Acquired lock "refresh_cache-268e5249-0156-45dd-bbbd-10105d16f269" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.169845] env[62952]: DEBUG nova.network.neutron [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 761.331996] env[62952]: INFO nova.compute.manager [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] [instance: 4bc891c1-e7e1-4196-9770-9bb42ed7499c] Took 1.02 seconds to deallocate network for instance. [ 761.518961] env[62952]: DEBUG nova.network.neutron [-] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.691948] env[62952]: DEBUG nova.network.neutron [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 761.772848] env[62952]: DEBUG nova.network.neutron [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.025057] env[62952]: INFO nova.compute.manager [-] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Took 1.03 seconds to deallocate network for instance. [ 762.025928] env[62952]: DEBUG nova.compute.claims [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 762.026347] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.281633] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Releasing lock "refresh_cache-268e5249-0156-45dd-bbbd-10105d16f269" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.281633] env[62952]: DEBUG nova.compute.manager [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 762.281633] env[62952]: DEBUG nova.compute.manager [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.281633] env[62952]: DEBUG nova.network.neutron [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 762.301443] env[62952]: DEBUG nova.network.neutron [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 762.360114] env[62952]: INFO nova.scheduler.client.report [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Deleted allocations for instance 4bc891c1-e7e1-4196-9770-9bb42ed7499c [ 762.553114] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f41d18-3d4d-463e-8f9f-47419ffbd4b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.561564] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9eb123-17f5-407a-acfc-6fc52f3e4851 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.597737] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387f33cb-6b25-43cd-88a1-cc67eb92d852 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.606171] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1b7f63-09d4-4c19-838a-aebb6bc7bed6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.621505] env[62952]: DEBUG nova.compute.provider_tree [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.805414] env[62952]: DEBUG nova.network.neutron [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.868599] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7472bed0-8ffc-4575-9c27-886c0d1c88bb tempest-ServersV294TestFqdnHostnames-238845146 tempest-ServersV294TestFqdnHostnames-238845146-project-member] Lock "4bc891c1-e7e1-4196-9770-9bb42ed7499c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.681s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.126568] env[62952]: DEBUG nova.scheduler.client.report [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.306430] env[62952]: INFO nova.compute.manager [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] [instance: 268e5249-0156-45dd-bbbd-10105d16f269] Took 1.03 seconds to deallocate network for instance. [ 763.371688] env[62952]: DEBUG nova.compute.manager [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 763.632654] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.467s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.632654] env[62952]: DEBUG nova.compute.manager [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 763.634945] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.512s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.893479] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.142018] env[62952]: DEBUG nova.compute.utils [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 764.144689] env[62952]: DEBUG nova.compute.manager [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 764.145156] env[62952]: DEBUG nova.network.neutron [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 764.193068] env[62952]: DEBUG nova.policy [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73138ce58f5c4bb1ac191d2dbd6328a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9bb3a6cd8d2d4f1f83c5c1558b2b5c37', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 764.347616] env[62952]: INFO nova.scheduler.client.report [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Deleted allocations for instance 268e5249-0156-45dd-bbbd-10105d16f269 [ 764.514616] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa1ef88a-d0a8-4930-ab75-bd206ef88dd1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.524072] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3978a9f5-eec3-44d9-afac-1b2536c29280 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.531872] env[62952]: DEBUG nova.network.neutron [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Successfully created port: 672ecb52-afb2-4f75-8227-ca1b0e7c21a1 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 764.560494] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e384f8d-765a-4696-9620-c3b25617e225 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.568502] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ff7d07-407e-4506-aa65-c8df7c109558 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.582333] env[62952]: DEBUG nova.compute.provider_tree [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.646027] env[62952]: DEBUG nova.compute.manager [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 764.858039] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b079884-ddbb-4813-bbbd-23a525315756 tempest-ServersNegativeTestMultiTenantJSON-2064021937 tempest-ServersNegativeTestMultiTenantJSON-2064021937-project-member] Lock "268e5249-0156-45dd-bbbd-10105d16f269" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.299s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.088164] env[62952]: DEBUG nova.scheduler.client.report [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 765.304170] env[62952]: DEBUG nova.compute.manager [req-7208b13c-4960-46cd-b6f1-f9773d208d02 req-04c9caf7-f076-464f-84cf-6d5e7abb7fae service nova] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Received event network-changed-672ecb52-afb2-4f75-8227-ca1b0e7c21a1 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 765.304652] env[62952]: DEBUG nova.compute.manager [req-7208b13c-4960-46cd-b6f1-f9773d208d02 req-04c9caf7-f076-464f-84cf-6d5e7abb7fae service nova] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Refreshing instance network info cache due to event network-changed-672ecb52-afb2-4f75-8227-ca1b0e7c21a1. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 765.304652] env[62952]: DEBUG oslo_concurrency.lockutils [req-7208b13c-4960-46cd-b6f1-f9773d208d02 req-04c9caf7-f076-464f-84cf-6d5e7abb7fae service nova] Acquiring lock "refresh_cache-a458a658-96f3-41b2-a9ee-2a6dd52bfeb4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.304652] env[62952]: DEBUG oslo_concurrency.lockutils [req-7208b13c-4960-46cd-b6f1-f9773d208d02 req-04c9caf7-f076-464f-84cf-6d5e7abb7fae service nova] Acquired lock "refresh_cache-a458a658-96f3-41b2-a9ee-2a6dd52bfeb4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.304783] env[62952]: DEBUG nova.network.neutron [req-7208b13c-4960-46cd-b6f1-f9773d208d02 req-04c9caf7-f076-464f-84cf-6d5e7abb7fae service nova] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Refreshing network info cache for port 672ecb52-afb2-4f75-8227-ca1b0e7c21a1 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 765.360542] env[62952]: DEBUG nova.compute.manager [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 765.420445] env[62952]: ERROR nova.compute.manager [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 672ecb52-afb2-4f75-8227-ca1b0e7c21a1, please check neutron logs for more information. [ 765.420445] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 765.420445] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.420445] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 765.420445] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 765.420445] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 765.420445] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 765.420445] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 765.420445] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.420445] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 765.420445] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.420445] env[62952]: ERROR nova.compute.manager raise self.value [ 765.420445] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 765.420445] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 765.420445] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.420445] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 765.423075] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.423075] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 765.423075] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 672ecb52-afb2-4f75-8227-ca1b0e7c21a1, please check neutron logs for more information. [ 765.423075] env[62952]: ERROR nova.compute.manager [ 765.423075] env[62952]: Traceback (most recent call last): [ 765.423075] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 765.423075] env[62952]: listener.cb(fileno) [ 765.423075] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.423075] env[62952]: result = function(*args, **kwargs) [ 765.423075] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 765.423075] env[62952]: return func(*args, **kwargs) [ 765.423075] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.423075] env[62952]: raise e [ 765.423075] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.423075] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 765.423075] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 765.423075] env[62952]: created_port_ids = self._update_ports_for_instance( [ 765.423075] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 765.423075] env[62952]: with excutils.save_and_reraise_exception(): [ 765.423075] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.423075] env[62952]: self.force_reraise() [ 765.423075] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.423075] env[62952]: raise self.value [ 765.423075] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 765.423075] env[62952]: updated_port = self._update_port( [ 765.423075] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.423075] env[62952]: _ensure_no_port_binding_failure(port) [ 765.423075] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.423075] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 765.423928] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 672ecb52-afb2-4f75-8227-ca1b0e7c21a1, please check neutron logs for more information. [ 765.423928] env[62952]: Removing descriptor: 19 [ 765.594020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.956s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.594020] env[62952]: ERROR nova.compute.manager [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9b472956-916e-4f25-a11f-fee191c954c1, please check neutron logs for more information. [ 765.594020] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Traceback (most recent call last): [ 765.594020] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 765.594020] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] self.driver.spawn(context, instance, image_meta, [ 765.594020] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 765.594020] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 765.594020] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 765.594020] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] vm_ref = self.build_virtual_machine(instance, [ 765.594358] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 765.594358] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] vif_infos = vmwarevif.get_vif_info(self._session, [ 765.594358] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 765.594358] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] for vif in network_info: [ 765.594358] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 765.594358] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] return self._sync_wrapper(fn, *args, **kwargs) [ 765.594358] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 765.594358] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] self.wait() [ 765.594358] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 765.594358] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] self[:] = self._gt.wait() [ 765.594358] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 765.594358] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] return self._exit_event.wait() [ 765.594358] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 765.594710] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] current.throw(*self._exc) [ 765.594710] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.594710] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] result = function(*args, **kwargs) [ 765.594710] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 765.594710] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] return func(*args, **kwargs) [ 765.594710] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.594710] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] raise e [ 765.594710] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.594710] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] nwinfo = self.network_api.allocate_for_instance( [ 765.594710] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 765.594710] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] created_port_ids = self._update_ports_for_instance( [ 765.594710] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 765.594710] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] with excutils.save_and_reraise_exception(): [ 765.595060] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.595060] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] self.force_reraise() [ 765.595060] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.595060] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] raise self.value [ 765.595060] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 765.595060] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] updated_port = self._update_port( [ 765.595060] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.595060] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] _ensure_no_port_binding_failure(port) [ 765.595060] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.595060] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] raise exception.PortBindingFailed(port_id=port['id']) [ 765.595060] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] nova.exception.PortBindingFailed: Binding failed for port 9b472956-916e-4f25-a11f-fee191c954c1, please check neutron logs for more information. [ 765.595060] env[62952]: ERROR nova.compute.manager [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] [ 765.595378] env[62952]: DEBUG nova.compute.utils [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Binding failed for port 9b472956-916e-4f25-a11f-fee191c954c1, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 765.595795] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.840s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.597549] env[62952]: INFO nova.compute.claims [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 765.602019] env[62952]: DEBUG nova.compute.manager [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Build of instance f9ed97a4-ec32-415b-958e-d362b622b1d4 was re-scheduled: Binding failed for port 9b472956-916e-4f25-a11f-fee191c954c1, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 765.602019] env[62952]: DEBUG nova.compute.manager [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 765.602019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "refresh_cache-f9ed97a4-ec32-415b-958e-d362b622b1d4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.602019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "refresh_cache-f9ed97a4-ec32-415b-958e-d362b622b1d4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.602225] env[62952]: DEBUG nova.network.neutron [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 765.662459] env[62952]: DEBUG nova.compute.manager [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 765.693976] env[62952]: DEBUG nova.virt.hardware [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 765.694616] env[62952]: DEBUG nova.virt.hardware [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 765.694616] env[62952]: DEBUG nova.virt.hardware [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 765.694760] env[62952]: DEBUG nova.virt.hardware [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 765.696369] env[62952]: DEBUG nova.virt.hardware [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 765.696369] env[62952]: DEBUG nova.virt.hardware [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 765.696369] env[62952]: DEBUG nova.virt.hardware [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 765.696369] env[62952]: DEBUG nova.virt.hardware [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 765.696369] env[62952]: DEBUG nova.virt.hardware [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 765.697543] env[62952]: DEBUG nova.virt.hardware [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 765.697543] env[62952]: DEBUG nova.virt.hardware [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 765.697543] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef4a541-9bdf-4d45-b38f-ab05aba135af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.705318] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d53427f-c8e5-43a3-a9f2-e1f4bf4b1012 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.721444] env[62952]: ERROR nova.compute.manager [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 672ecb52-afb2-4f75-8227-ca1b0e7c21a1, please check neutron logs for more information. [ 765.721444] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Traceback (most recent call last): [ 765.721444] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 765.721444] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] yield resources [ 765.721444] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 765.721444] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] self.driver.spawn(context, instance, image_meta, [ 765.721444] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 765.721444] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 765.721444] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 765.721444] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] vm_ref = self.build_virtual_machine(instance, [ 765.721444] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 765.722225] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] vif_infos = vmwarevif.get_vif_info(self._session, [ 765.722225] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 765.722225] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] for vif in network_info: [ 765.722225] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 765.722225] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] return self._sync_wrapper(fn, *args, **kwargs) [ 765.722225] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 765.722225] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] self.wait() [ 765.722225] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 765.722225] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] self[:] = self._gt.wait() [ 765.722225] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 765.722225] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] return self._exit_event.wait() [ 765.722225] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 765.722225] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] current.throw(*self._exc) [ 765.722599] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.722599] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] result = function(*args, **kwargs) [ 765.722599] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 765.722599] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] return func(*args, **kwargs) [ 765.722599] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.722599] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] raise e [ 765.722599] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.722599] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] nwinfo = self.network_api.allocate_for_instance( [ 765.722599] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 765.722599] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] created_port_ids = self._update_ports_for_instance( [ 765.722599] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 765.722599] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] with excutils.save_and_reraise_exception(): [ 765.722599] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.722981] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] self.force_reraise() [ 765.722981] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.722981] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] raise self.value [ 765.722981] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 765.722981] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] updated_port = self._update_port( [ 765.722981] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.722981] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] _ensure_no_port_binding_failure(port) [ 765.722981] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.722981] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] raise exception.PortBindingFailed(port_id=port['id']) [ 765.722981] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] nova.exception.PortBindingFailed: Binding failed for port 672ecb52-afb2-4f75-8227-ca1b0e7c21a1, please check neutron logs for more information. [ 765.722981] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] [ 765.722981] env[62952]: INFO nova.compute.manager [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Terminating instance [ 765.723858] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Acquiring lock "refresh_cache-a458a658-96f3-41b2-a9ee-2a6dd52bfeb4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.824628] env[62952]: DEBUG nova.network.neutron [req-7208b13c-4960-46cd-b6f1-f9773d208d02 req-04c9caf7-f076-464f-84cf-6d5e7abb7fae service nova] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 765.888104] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.914025] env[62952]: DEBUG nova.network.neutron [req-7208b13c-4960-46cd-b6f1-f9773d208d02 req-04c9caf7-f076-464f-84cf-6d5e7abb7fae service nova] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.126859] env[62952]: DEBUG nova.network.neutron [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 766.216488] env[62952]: DEBUG nova.network.neutron [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.417373] env[62952]: DEBUG oslo_concurrency.lockutils [req-7208b13c-4960-46cd-b6f1-f9773d208d02 req-04c9caf7-f076-464f-84cf-6d5e7abb7fae service nova] Releasing lock "refresh_cache-a458a658-96f3-41b2-a9ee-2a6dd52bfeb4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.417898] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Acquired lock "refresh_cache-a458a658-96f3-41b2-a9ee-2a6dd52bfeb4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.422895] env[62952]: DEBUG nova.network.neutron [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 766.718487] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "refresh_cache-f9ed97a4-ec32-415b-958e-d362b622b1d4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.718790] env[62952]: DEBUG nova.compute.manager [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 766.718986] env[62952]: DEBUG nova.compute.manager [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.719468] env[62952]: DEBUG nova.network.neutron [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 766.741252] env[62952]: DEBUG nova.network.neutron [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 766.948680] env[62952]: DEBUG nova.network.neutron [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 767.007366] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-273a4614-ac19-4468-96fc-18d8c4685650 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.015210] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac78131-b83e-44eb-abc9-f83be4674272 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.047634] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac01e6d-fa6c-4fdc-884d-a02537d79b69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.055006] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c37c1e5-3a2d-4222-aea3-c30f305711a7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.069051] env[62952]: DEBUG nova.compute.provider_tree [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.113949] env[62952]: DEBUG nova.network.neutron [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.248786] env[62952]: DEBUG nova.network.neutron [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.440970] env[62952]: DEBUG nova.compute.manager [req-47d848cf-9a6f-456b-86ec-2126f6221925 req-0571aa02-2d6a-4867-b39a-595d5ce3bf9b service nova] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Received event network-vif-deleted-672ecb52-afb2-4f75-8227-ca1b0e7c21a1 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 767.572510] env[62952]: DEBUG nova.scheduler.client.report [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 767.616470] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Releasing lock "refresh_cache-a458a658-96f3-41b2-a9ee-2a6dd52bfeb4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.616893] env[62952]: DEBUG nova.compute.manager [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 767.617097] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 767.617388] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d270c60d-0a86-4e9f-b461-91162ec53074 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.627911] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73fd8aa1-f5d9-4d4e-ae83-6519d732d465 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.650909] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a458a658-96f3-41b2-a9ee-2a6dd52bfeb4 could not be found. [ 767.651304] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 767.651393] env[62952]: INFO nova.compute.manager [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 767.651554] env[62952]: DEBUG oslo.service.loopingcall [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 767.651771] env[62952]: DEBUG nova.compute.manager [-] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 767.651864] env[62952]: DEBUG nova.network.neutron [-] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 767.673072] env[62952]: DEBUG nova.network.neutron [-] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 767.750966] env[62952]: INFO nova.compute.manager [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: f9ed97a4-ec32-415b-958e-d362b622b1d4] Took 1.03 seconds to deallocate network for instance. [ 768.077046] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.078684] env[62952]: DEBUG nova.compute.manager [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 768.081571] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.315s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.086087] env[62952]: INFO nova.compute.claims [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 768.175722] env[62952]: DEBUG nova.network.neutron [-] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.590245] env[62952]: DEBUG nova.compute.utils [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 768.593573] env[62952]: DEBUG nova.compute.manager [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 768.593862] env[62952]: DEBUG nova.network.neutron [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 768.655259] env[62952]: DEBUG nova.policy [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c5cab9fec6604a0881d677df17dda195', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c51c89d19ed4111a5c1631f0ad47cba', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 768.678234] env[62952]: INFO nova.compute.manager [-] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Took 1.03 seconds to deallocate network for instance. [ 768.683888] env[62952]: DEBUG nova.compute.claims [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 768.684096] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.799422] env[62952]: INFO nova.scheduler.client.report [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Deleted allocations for instance f9ed97a4-ec32-415b-958e-d362b622b1d4 [ 769.036445] env[62952]: DEBUG nova.network.neutron [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Successfully created port: b3154985-2dd0-4f1e-b7fe-d36d43a66c89 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 769.094146] env[62952]: DEBUG nova.compute.manager [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 769.315686] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ffa6195-0886-4dd2-abd8-c84152fb5be5 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "f9ed97a4-ec32-415b-958e-d362b622b1d4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.754s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.420800] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67df9f1-71b6-49c2-a381-a5847657c598 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.429139] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f9bfb3-4ada-4236-bc88-627a4298e26c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.461578] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3262f587-ebab-4c96-925a-cfb0d74952b6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.469017] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86efd933-e52c-4aa4-8b9e-74e8bd600323 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.482839] env[62952]: DEBUG nova.compute.provider_tree [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.818258] env[62952]: DEBUG nova.compute.manager [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.847377] env[62952]: DEBUG nova.compute.manager [req-015eea43-5860-43ed-b39b-01d1eb942c5f req-bca450f0-543e-4530-9334-d8990093fa19 service nova] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Received event network-changed-b3154985-2dd0-4f1e-b7fe-d36d43a66c89 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 769.847580] env[62952]: DEBUG nova.compute.manager [req-015eea43-5860-43ed-b39b-01d1eb942c5f req-bca450f0-543e-4530-9334-d8990093fa19 service nova] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Refreshing instance network info cache due to event network-changed-b3154985-2dd0-4f1e-b7fe-d36d43a66c89. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 769.847795] env[62952]: DEBUG oslo_concurrency.lockutils [req-015eea43-5860-43ed-b39b-01d1eb942c5f req-bca450f0-543e-4530-9334-d8990093fa19 service nova] Acquiring lock "refresh_cache-dd88d70d-17a3-4059-9092-5d6cec8b1ae4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.847975] env[62952]: DEBUG oslo_concurrency.lockutils [req-015eea43-5860-43ed-b39b-01d1eb942c5f req-bca450f0-543e-4530-9334-d8990093fa19 service nova] Acquired lock "refresh_cache-dd88d70d-17a3-4059-9092-5d6cec8b1ae4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.848133] env[62952]: DEBUG nova.network.neutron [req-015eea43-5860-43ed-b39b-01d1eb942c5f req-bca450f0-543e-4530-9334-d8990093fa19 service nova] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Refreshing network info cache for port b3154985-2dd0-4f1e-b7fe-d36d43a66c89 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 769.987430] env[62952]: DEBUG nova.scheduler.client.report [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 770.012373] env[62952]: ERROR nova.compute.manager [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b3154985-2dd0-4f1e-b7fe-d36d43a66c89, please check neutron logs for more information. [ 770.012373] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 770.012373] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.012373] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 770.012373] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 770.012373] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 770.012373] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 770.012373] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 770.012373] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.012373] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 770.012373] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.012373] env[62952]: ERROR nova.compute.manager raise self.value [ 770.012373] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 770.012373] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 770.012373] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.012373] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 770.012823] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.012823] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 770.012823] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b3154985-2dd0-4f1e-b7fe-d36d43a66c89, please check neutron logs for more information. [ 770.012823] env[62952]: ERROR nova.compute.manager [ 770.012823] env[62952]: Traceback (most recent call last): [ 770.012823] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 770.012823] env[62952]: listener.cb(fileno) [ 770.012823] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 770.012823] env[62952]: result = function(*args, **kwargs) [ 770.012823] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 770.012823] env[62952]: return func(*args, **kwargs) [ 770.012823] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 770.012823] env[62952]: raise e [ 770.012823] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.012823] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 770.012823] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 770.012823] env[62952]: created_port_ids = self._update_ports_for_instance( [ 770.012823] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 770.012823] env[62952]: with excutils.save_and_reraise_exception(): [ 770.012823] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.012823] env[62952]: self.force_reraise() [ 770.012823] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.012823] env[62952]: raise self.value [ 770.012823] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 770.012823] env[62952]: updated_port = self._update_port( [ 770.012823] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.012823] env[62952]: _ensure_no_port_binding_failure(port) [ 770.012823] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.012823] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 770.013623] env[62952]: nova.exception.PortBindingFailed: Binding failed for port b3154985-2dd0-4f1e-b7fe-d36d43a66c89, please check neutron logs for more information. [ 770.013623] env[62952]: Removing descriptor: 19 [ 770.111331] env[62952]: DEBUG nova.compute.manager [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.138335] env[62952]: DEBUG nova.virt.hardware [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.138625] env[62952]: DEBUG nova.virt.hardware [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.138787] env[62952]: DEBUG nova.virt.hardware [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.138969] env[62952]: DEBUG nova.virt.hardware [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.139175] env[62952]: DEBUG nova.virt.hardware [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.139345] env[62952]: DEBUG nova.virt.hardware [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.139557] env[62952]: DEBUG nova.virt.hardware [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.139713] env[62952]: DEBUG nova.virt.hardware [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.139872] env[62952]: DEBUG nova.virt.hardware [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.140042] env[62952]: DEBUG nova.virt.hardware [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.140220] env[62952]: DEBUG nova.virt.hardware [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.141525] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e3bfe1-70ad-48dd-a1c9-c5fff41ce3a0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.150714] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385159d0-e112-4055-a812-8c31e18c91ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.167122] env[62952]: ERROR nova.compute.manager [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b3154985-2dd0-4f1e-b7fe-d36d43a66c89, please check neutron logs for more information. [ 770.167122] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Traceback (most recent call last): [ 770.167122] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 770.167122] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] yield resources [ 770.167122] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 770.167122] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] self.driver.spawn(context, instance, image_meta, [ 770.167122] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 770.167122] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 770.167122] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 770.167122] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] vm_ref = self.build_virtual_machine(instance, [ 770.167122] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 770.167576] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] vif_infos = vmwarevif.get_vif_info(self._session, [ 770.167576] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 770.167576] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] for vif in network_info: [ 770.167576] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 770.167576] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] return self._sync_wrapper(fn, *args, **kwargs) [ 770.167576] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 770.167576] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] self.wait() [ 770.167576] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 770.167576] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] self[:] = self._gt.wait() [ 770.167576] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 770.167576] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] return self._exit_event.wait() [ 770.167576] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 770.167576] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] current.throw(*self._exc) [ 770.168412] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 770.168412] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] result = function(*args, **kwargs) [ 770.168412] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 770.168412] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] return func(*args, **kwargs) [ 770.168412] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 770.168412] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] raise e [ 770.168412] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.168412] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] nwinfo = self.network_api.allocate_for_instance( [ 770.168412] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 770.168412] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] created_port_ids = self._update_ports_for_instance( [ 770.168412] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 770.168412] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] with excutils.save_and_reraise_exception(): [ 770.168412] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.168727] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] self.force_reraise() [ 770.168727] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.168727] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] raise self.value [ 770.168727] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 770.168727] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] updated_port = self._update_port( [ 770.168727] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.168727] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] _ensure_no_port_binding_failure(port) [ 770.168727] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.168727] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] raise exception.PortBindingFailed(port_id=port['id']) [ 770.168727] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] nova.exception.PortBindingFailed: Binding failed for port b3154985-2dd0-4f1e-b7fe-d36d43a66c89, please check neutron logs for more information. [ 770.168727] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] [ 770.168727] env[62952]: INFO nova.compute.manager [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Terminating instance [ 770.169082] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquiring lock "refresh_cache-dd88d70d-17a3-4059-9092-5d6cec8b1ae4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.349794] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.369821] env[62952]: DEBUG nova.network.neutron [req-015eea43-5860-43ed-b39b-01d1eb942c5f req-bca450f0-543e-4530-9334-d8990093fa19 service nova] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 770.438432] env[62952]: DEBUG nova.network.neutron [req-015eea43-5860-43ed-b39b-01d1eb942c5f req-bca450f0-543e-4530-9334-d8990093fa19 service nova] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.491050] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.409s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.491619] env[62952]: DEBUG nova.compute.manager [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 770.494557] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.891s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.496636] env[62952]: INFO nova.compute.claims [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 770.941148] env[62952]: DEBUG oslo_concurrency.lockutils [req-015eea43-5860-43ed-b39b-01d1eb942c5f req-bca450f0-543e-4530-9334-d8990093fa19 service nova] Releasing lock "refresh_cache-dd88d70d-17a3-4059-9092-5d6cec8b1ae4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.941578] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquired lock "refresh_cache-dd88d70d-17a3-4059-9092-5d6cec8b1ae4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.941761] env[62952]: DEBUG nova.network.neutron [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 771.001176] env[62952]: DEBUG nova.compute.utils [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 771.004391] env[62952]: DEBUG nova.compute.manager [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Not allocating networking since 'none' was specified. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 771.459131] env[62952]: DEBUG nova.network.neutron [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 771.505677] env[62952]: DEBUG nova.compute.manager [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 771.541247] env[62952]: DEBUG nova.network.neutron [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.786069] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a04c84e-9c5c-496f-a031-fd31b8166fb9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.793731] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963253eb-eac5-4949-a073-dffb1978b73e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.823330] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f3ac4d-7205-443f-ab9d-f213e2a5c344 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.830052] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68925c4f-2810-4470-b696-77632af5784a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.842571] env[62952]: DEBUG nova.compute.provider_tree [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.874718] env[62952]: DEBUG nova.compute.manager [req-2f58c230-bbb9-4fdf-90c1-e53f673b47e4 req-8cdf5c91-50e5-4158-b459-c332f2bb2fa8 service nova] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Received event network-vif-deleted-b3154985-2dd0-4f1e-b7fe-d36d43a66c89 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 772.010460] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.010700] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.042642] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Releasing lock "refresh_cache-dd88d70d-17a3-4059-9092-5d6cec8b1ae4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.042946] env[62952]: DEBUG nova.compute.manager [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 772.043155] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 772.044132] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dcb24ebe-dca7-4f0d-9a66-c749035453e0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.053280] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61d859bf-68f3-4dbc-b665-75ac845df5eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.075140] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dd88d70d-17a3-4059-9092-5d6cec8b1ae4 could not be found. [ 772.075367] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 772.075784] env[62952]: INFO nova.compute.manager [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 772.075784] env[62952]: DEBUG oslo.service.loopingcall [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.075969] env[62952]: DEBUG nova.compute.manager [-] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 772.076100] env[62952]: DEBUG nova.network.neutron [-] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 772.093673] env[62952]: DEBUG nova.network.neutron [-] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 772.345411] env[62952]: DEBUG nova.scheduler.client.report [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 772.516706] env[62952]: DEBUG nova.compute.manager [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 772.540967] env[62952]: DEBUG nova.virt.hardware [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 772.541221] env[62952]: DEBUG nova.virt.hardware [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 772.541379] env[62952]: DEBUG nova.virt.hardware [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 772.541556] env[62952]: DEBUG nova.virt.hardware [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 772.541700] env[62952]: DEBUG nova.virt.hardware [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 772.541839] env[62952]: DEBUG nova.virt.hardware [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 772.542116] env[62952]: DEBUG nova.virt.hardware [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 772.542292] env[62952]: DEBUG nova.virt.hardware [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 772.542458] env[62952]: DEBUG nova.virt.hardware [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 772.542657] env[62952]: DEBUG nova.virt.hardware [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 772.542836] env[62952]: DEBUG nova.virt.hardware [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 772.543728] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f534e58-55a8-4ed8-afa9-97eea6cab4ba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.552230] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7ce134-75b1-4f1d-bcb4-1b7d94718abb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.565621] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 772.571342] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Creating folder: Project (77603063745e4ecd945d040a29465106). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 772.571627] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e932e437-430e-4af3-b2ae-c9491d592f2a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.583752] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Created folder: Project (77603063745e4ecd945d040a29465106) in parent group-v271811. [ 772.583937] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Creating folder: Instances. Parent ref: group-v271826. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 772.584176] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f97c5c0-b255-45a6-93ba-7623fa619658 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.594080] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Created folder: Instances in parent group-v271826. [ 772.594080] env[62952]: DEBUG oslo.service.loopingcall [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.594080] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 772.594080] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7f338425-7d35-4ef2-bbeb-8837ef1e6492 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.605547] env[62952]: DEBUG nova.network.neutron [-] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.610676] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 772.610676] env[62952]: value = "task-1263102" [ 772.610676] env[62952]: _type = "Task" [ 772.610676] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.619163] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263102, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.851110] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.356s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.851384] env[62952]: DEBUG nova.compute.manager [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 772.853900] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.253s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.855433] env[62952]: INFO nova.compute.claims [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 773.108304] env[62952]: INFO nova.compute.manager [-] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Took 1.03 seconds to deallocate network for instance. [ 773.110610] env[62952]: DEBUG nova.compute.claims [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 773.110789] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.119899] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263102, 'name': CreateVM_Task, 'duration_secs': 0.237834} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.120060] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 773.120436] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.120616] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.120917] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 773.121168] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a82251b-1dd5-4392-83bc-6be2498186ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.125432] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 773.125432] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b749ee-55d8-6dbf-eace-58838cf932a6" [ 773.125432] env[62952]: _type = "Task" [ 773.125432] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.132678] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b749ee-55d8-6dbf-eace-58838cf932a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.360284] env[62952]: DEBUG nova.compute.utils [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 773.363590] env[62952]: DEBUG nova.compute.manager [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 773.363763] env[62952]: DEBUG nova.network.neutron [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 773.413931] env[62952]: DEBUG nova.policy [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7eadc05f193645e6b6e86e9cdbbbaa71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '076fb3c53420447caf047f8422435152', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 773.637813] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b749ee-55d8-6dbf-eace-58838cf932a6, 'name': SearchDatastore_Task, 'duration_secs': 0.01111} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.637813] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.637813] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 773.637813] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.637962] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.637962] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 773.637962] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-03155e26-f26d-4270-b51f-8e64de4146d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.645523] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 773.645699] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 773.646413] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8278298-cbcd-491b-99c6-54f361c282c1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.651551] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 773.651551] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d7d5e0-26fc-7b8a-25b4-e891595b2aa8" [ 773.651551] env[62952]: _type = "Task" [ 773.651551] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.659175] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d7d5e0-26fc-7b8a-25b4-e891595b2aa8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.669685] env[62952]: DEBUG nova.network.neutron [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Successfully created port: 0613de0a-f8a0-4c51-a045-3442c42f143a {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 773.864747] env[62952]: DEBUG nova.compute.manager [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 774.162589] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d7d5e0-26fc-7b8a-25b4-e891595b2aa8, 'name': SearchDatastore_Task, 'duration_secs': 0.008438} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.166140] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6dae8676-be97-48cb-bf24-26eada912730 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.175322] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 774.175322] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]522afe9c-500c-67b6-703a-6d6bb00ac027" [ 774.175322] env[62952]: _type = "Task" [ 774.175322] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.184733] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]522afe9c-500c-67b6-703a-6d6bb00ac027, 'name': SearchDatastore_Task, 'duration_secs': 0.008519} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.184858] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.185864] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 472e60f1-9b6c-415e-8a92-19235ad2e547/472e60f1-9b6c-415e-8a92-19235ad2e547.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 774.185864] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b0d804e5-81b1-4cc4-9d00-69d5ab51be07 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.192459] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 774.192459] env[62952]: value = "task-1263103" [ 774.192459] env[62952]: _type = "Task" [ 774.192459] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.200038] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263103, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.262636] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b50a65d-4fb8-4ff3-97b3-1c8216eaec21 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.270061] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f687765f-6f7b-462a-b875-42b4169c08b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.300907] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37136e6d-71e2-4a85-80e5-67e87116ef2e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.308273] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee0dfaf4-5dfd-4528-bf99-02ce7effd4ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.321117] env[62952]: DEBUG nova.compute.provider_tree [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.345329] env[62952]: DEBUG nova.compute.manager [req-1a9ce3ce-3902-4d53-aafd-1642d0eca598 req-e526efd1-6f2f-4eaf-a5f2-0dddfc7e45c1 service nova] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Received event network-changed-0613de0a-f8a0-4c51-a045-3442c42f143a {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 774.345507] env[62952]: DEBUG nova.compute.manager [req-1a9ce3ce-3902-4d53-aafd-1642d0eca598 req-e526efd1-6f2f-4eaf-a5f2-0dddfc7e45c1 service nova] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Refreshing instance network info cache due to event network-changed-0613de0a-f8a0-4c51-a045-3442c42f143a. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 774.345732] env[62952]: DEBUG oslo_concurrency.lockutils [req-1a9ce3ce-3902-4d53-aafd-1642d0eca598 req-e526efd1-6f2f-4eaf-a5f2-0dddfc7e45c1 service nova] Acquiring lock "refresh_cache-b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.345871] env[62952]: DEBUG oslo_concurrency.lockutils [req-1a9ce3ce-3902-4d53-aafd-1642d0eca598 req-e526efd1-6f2f-4eaf-a5f2-0dddfc7e45c1 service nova] Acquired lock "refresh_cache-b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.346973] env[62952]: DEBUG nova.network.neutron [req-1a9ce3ce-3902-4d53-aafd-1642d0eca598 req-e526efd1-6f2f-4eaf-a5f2-0dddfc7e45c1 service nova] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Refreshing network info cache for port 0613de0a-f8a0-4c51-a045-3442c42f143a {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 774.541619] env[62952]: ERROR nova.compute.manager [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0613de0a-f8a0-4c51-a045-3442c42f143a, please check neutron logs for more information. [ 774.541619] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 774.541619] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 774.541619] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 774.541619] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 774.541619] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 774.541619] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 774.541619] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 774.541619] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 774.541619] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 774.541619] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 774.541619] env[62952]: ERROR nova.compute.manager raise self.value [ 774.541619] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 774.541619] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 774.541619] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 774.541619] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 774.542233] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 774.542233] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 774.542233] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0613de0a-f8a0-4c51-a045-3442c42f143a, please check neutron logs for more information. [ 774.542233] env[62952]: ERROR nova.compute.manager [ 774.542233] env[62952]: Traceback (most recent call last): [ 774.542233] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 774.542233] env[62952]: listener.cb(fileno) [ 774.542233] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 774.542233] env[62952]: result = function(*args, **kwargs) [ 774.542233] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 774.542233] env[62952]: return func(*args, **kwargs) [ 774.542233] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 774.542233] env[62952]: raise e [ 774.542233] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 774.542233] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 774.542233] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 774.542233] env[62952]: created_port_ids = self._update_ports_for_instance( [ 774.542233] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 774.542233] env[62952]: with excutils.save_and_reraise_exception(): [ 774.542233] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 774.542233] env[62952]: self.force_reraise() [ 774.542233] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 774.542233] env[62952]: raise self.value [ 774.542233] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 774.542233] env[62952]: updated_port = self._update_port( [ 774.542233] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 774.542233] env[62952]: _ensure_no_port_binding_failure(port) [ 774.542233] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 774.542233] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 774.543412] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 0613de0a-f8a0-4c51-a045-3442c42f143a, please check neutron logs for more information. [ 774.543412] env[62952]: Removing descriptor: 19 [ 774.702610] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263103, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462842} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.702610] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 472e60f1-9b6c-415e-8a92-19235ad2e547/472e60f1-9b6c-415e-8a92-19235ad2e547.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 774.702988] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 774.703231] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eeaf8c02-cfec-4cec-ad10-2273bc34c46a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.709547] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 774.709547] env[62952]: value = "task-1263104" [ 774.709547] env[62952]: _type = "Task" [ 774.709547] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.717138] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263104, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.825496] env[62952]: DEBUG nova.scheduler.client.report [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.863387] env[62952]: DEBUG nova.network.neutron [req-1a9ce3ce-3902-4d53-aafd-1642d0eca598 req-e526efd1-6f2f-4eaf-a5f2-0dddfc7e45c1 service nova] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 774.878933] env[62952]: DEBUG nova.compute.manager [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 774.903455] env[62952]: DEBUG nova.virt.hardware [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 774.903710] env[62952]: DEBUG nova.virt.hardware [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 774.903865] env[62952]: DEBUG nova.virt.hardware [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 774.904057] env[62952]: DEBUG nova.virt.hardware [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 774.904206] env[62952]: DEBUG nova.virt.hardware [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 774.904351] env[62952]: DEBUG nova.virt.hardware [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 774.904546] env[62952]: DEBUG nova.virt.hardware [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 774.904698] env[62952]: DEBUG nova.virt.hardware [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 774.904858] env[62952]: DEBUG nova.virt.hardware [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 774.905023] env[62952]: DEBUG nova.virt.hardware [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 774.905200] env[62952]: DEBUG nova.virt.hardware [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 774.906364] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-379a3623-c888-455e-bb67-a564f0733a70 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.915023] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e365b85d-3935-4f74-aa01-b9fdb0f141a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.928150] env[62952]: ERROR nova.compute.manager [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0613de0a-f8a0-4c51-a045-3442c42f143a, please check neutron logs for more information. [ 774.928150] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Traceback (most recent call last): [ 774.928150] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 774.928150] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] yield resources [ 774.928150] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 774.928150] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] self.driver.spawn(context, instance, image_meta, [ 774.928150] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 774.928150] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 774.928150] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 774.928150] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] vm_ref = self.build_virtual_machine(instance, [ 774.928150] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 774.928549] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 774.928549] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 774.928549] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] for vif in network_info: [ 774.928549] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 774.928549] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] return self._sync_wrapper(fn, *args, **kwargs) [ 774.928549] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 774.928549] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] self.wait() [ 774.928549] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 774.928549] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] self[:] = self._gt.wait() [ 774.928549] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 774.928549] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] return self._exit_event.wait() [ 774.928549] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 774.928549] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] current.throw(*self._exc) [ 774.928929] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 774.928929] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] result = function(*args, **kwargs) [ 774.928929] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 774.928929] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] return func(*args, **kwargs) [ 774.928929] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 774.928929] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] raise e [ 774.928929] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 774.928929] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] nwinfo = self.network_api.allocate_for_instance( [ 774.928929] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 774.928929] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] created_port_ids = self._update_ports_for_instance( [ 774.928929] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 774.928929] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] with excutils.save_and_reraise_exception(): [ 774.928929] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 774.929289] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] self.force_reraise() [ 774.929289] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 774.929289] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] raise self.value [ 774.929289] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 774.929289] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] updated_port = self._update_port( [ 774.929289] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 774.929289] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] _ensure_no_port_binding_failure(port) [ 774.929289] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 774.929289] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] raise exception.PortBindingFailed(port_id=port['id']) [ 774.929289] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] nova.exception.PortBindingFailed: Binding failed for port 0613de0a-f8a0-4c51-a045-3442c42f143a, please check neutron logs for more information. [ 774.929289] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] [ 774.929289] env[62952]: INFO nova.compute.manager [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Terminating instance [ 774.930400] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "refresh_cache-b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.947456] env[62952]: DEBUG nova.network.neutron [req-1a9ce3ce-3902-4d53-aafd-1642d0eca598 req-e526efd1-6f2f-4eaf-a5f2-0dddfc7e45c1 service nova] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.219225] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263104, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062048} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.219509] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 775.220248] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96874d2a-d902-4597-a5fa-78d1a994e606 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.239014] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Reconfiguring VM instance instance-00000024 to attach disk [datastore1] 472e60f1-9b6c-415e-8a92-19235ad2e547/472e60f1-9b6c-415e-8a92-19235ad2e547.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 775.239263] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2678df1a-a101-4cfc-8d67-156cc6ce66f4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.258309] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 775.258309] env[62952]: value = "task-1263105" [ 775.258309] env[62952]: _type = "Task" [ 775.258309] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.265584] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263105, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.330757] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.331310] env[62952]: DEBUG nova.compute.manager [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 775.333986] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.845s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.336063] env[62952]: INFO nova.compute.claims [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 775.450166] env[62952]: DEBUG oslo_concurrency.lockutils [req-1a9ce3ce-3902-4d53-aafd-1642d0eca598 req-e526efd1-6f2f-4eaf-a5f2-0dddfc7e45c1 service nova] Releasing lock "refresh_cache-b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.450645] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "refresh_cache-b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.450828] env[62952]: DEBUG nova.network.neutron [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 775.767984] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263105, 'name': ReconfigVM_Task, 'duration_secs': 0.261739} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.769297] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Reconfigured VM instance instance-00000024 to attach disk [datastore1] 472e60f1-9b6c-415e-8a92-19235ad2e547/472e60f1-9b6c-415e-8a92-19235ad2e547.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 775.769297] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a2a2d172-24a2-42a1-976b-619f1b7808c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.778033] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 775.778033] env[62952]: value = "task-1263106" [ 775.778033] env[62952]: _type = "Task" [ 775.778033] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.787439] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263106, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.839876] env[62952]: DEBUG nova.compute.utils [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 775.843229] env[62952]: DEBUG nova.compute.manager [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 775.843442] env[62952]: DEBUG nova.network.neutron [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 775.884451] env[62952]: DEBUG nova.policy [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd04160e7a69746f8b2e8672e462d632c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '174cc8c213104d3787d251b8483e624e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 775.971674] env[62952]: DEBUG nova.network.neutron [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 776.092587] env[62952]: DEBUG nova.network.neutron [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.194538] env[62952]: DEBUG nova.network.neutron [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Successfully created port: 52a0fa94-676e-4957-a12e-c9c1cd324b75 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 776.286726] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263106, 'name': Rename_Task, 'duration_secs': 0.13151} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.287028] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 776.287229] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c585c41b-d13e-4d17-ab8e-d87e78492899 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.294376] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 776.294376] env[62952]: value = "task-1263107" [ 776.294376] env[62952]: _type = "Task" [ 776.294376] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.302801] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263107, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.344634] env[62952]: DEBUG nova.compute.manager [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 776.374691] env[62952]: DEBUG nova.compute.manager [req-4751fc17-2b21-44e6-ab58-b3d0f6e5a2d1 req-975f81f0-ca83-4039-b8d0-4ffe1e97bac2 service nova] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Received event network-vif-deleted-0613de0a-f8a0-4c51-a045-3442c42f143a {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 776.595271] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "refresh_cache-b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.595725] env[62952]: DEBUG nova.compute.manager [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 776.595929] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 776.596254] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-52c4d10c-3817-4745-ba05-bb1770ee15c8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.611419] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e398a5d-8e5a-4430-814c-9bc71d66ffe0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.633759] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6 could not be found. [ 776.633994] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 776.634188] env[62952]: INFO nova.compute.manager [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 776.634428] env[62952]: DEBUG oslo.service.loopingcall [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 776.635035] env[62952]: DEBUG nova.compute.manager [-] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 776.635140] env[62952]: DEBUG nova.network.neutron [-] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 776.653927] env[62952]: DEBUG nova.network.neutron [-] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 776.699585] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89201685-d20b-435e-8c8f-67d530442e1d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.707808] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4389a08-002b-4e04-b41c-14a1a59be7e4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.744033] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e16d6633-60c3-42a8-a5ea-a9dfdf46af3d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.751579] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6b1b1cf-5187-418e-852d-ff74ba82db23 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.767554] env[62952]: DEBUG nova.compute.provider_tree [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.803535] env[62952]: DEBUG oslo_vmware.api [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263107, 'name': PowerOnVM_Task, 'duration_secs': 0.415076} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.804144] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 776.804144] env[62952]: INFO nova.compute.manager [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Took 4.29 seconds to spawn the instance on the hypervisor. [ 776.804144] env[62952]: DEBUG nova.compute.manager [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 776.804973] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0902f042-9ee2-4654-a331-3c7a0cda4dde {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.156706] env[62952]: DEBUG nova.network.neutron [-] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.170794] env[62952]: ERROR nova.compute.manager [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 52a0fa94-676e-4957-a12e-c9c1cd324b75, please check neutron logs for more information. [ 777.170794] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 777.170794] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.170794] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 777.170794] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 777.170794] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 777.170794] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 777.170794] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 777.170794] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.170794] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 777.170794] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.170794] env[62952]: ERROR nova.compute.manager raise self.value [ 777.170794] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 777.170794] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 777.170794] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.170794] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 777.171276] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.171276] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 777.171276] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 52a0fa94-676e-4957-a12e-c9c1cd324b75, please check neutron logs for more information. [ 777.171276] env[62952]: ERROR nova.compute.manager [ 777.171276] env[62952]: Traceback (most recent call last): [ 777.171276] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 777.171276] env[62952]: listener.cb(fileno) [ 777.171276] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 777.171276] env[62952]: result = function(*args, **kwargs) [ 777.171276] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 777.171276] env[62952]: return func(*args, **kwargs) [ 777.171276] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 777.171276] env[62952]: raise e [ 777.171276] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.171276] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 777.171276] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 777.171276] env[62952]: created_port_ids = self._update_ports_for_instance( [ 777.171276] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 777.171276] env[62952]: with excutils.save_and_reraise_exception(): [ 777.171276] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.171276] env[62952]: self.force_reraise() [ 777.171276] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.171276] env[62952]: raise self.value [ 777.171276] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 777.171276] env[62952]: updated_port = self._update_port( [ 777.171276] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.171276] env[62952]: _ensure_no_port_binding_failure(port) [ 777.171276] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.171276] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 777.171980] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 52a0fa94-676e-4957-a12e-c9c1cd324b75, please check neutron logs for more information. [ 777.171980] env[62952]: Removing descriptor: 19 [ 777.272536] env[62952]: DEBUG nova.scheduler.client.report [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 777.323460] env[62952]: INFO nova.compute.manager [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Took 38.61 seconds to build instance. [ 777.355604] env[62952]: DEBUG nova.compute.manager [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 777.381537] env[62952]: DEBUG nova.virt.hardware [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 777.381715] env[62952]: DEBUG nova.virt.hardware [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 777.381923] env[62952]: DEBUG nova.virt.hardware [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 777.382136] env[62952]: DEBUG nova.virt.hardware [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 777.382285] env[62952]: DEBUG nova.virt.hardware [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 777.382431] env[62952]: DEBUG nova.virt.hardware [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 777.382639] env[62952]: DEBUG nova.virt.hardware [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 777.382794] env[62952]: DEBUG nova.virt.hardware [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 777.382961] env[62952]: DEBUG nova.virt.hardware [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 777.383138] env[62952]: DEBUG nova.virt.hardware [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 777.383344] env[62952]: DEBUG nova.virt.hardware [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 777.384215] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f91c4c9f-837e-4e5f-a02f-0ebc2406e9f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.392496] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b86855cc-22cc-4f6c-a3c6-656202d9c851 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.408329] env[62952]: ERROR nova.compute.manager [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 52a0fa94-676e-4957-a12e-c9c1cd324b75, please check neutron logs for more information. [ 777.408329] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Traceback (most recent call last): [ 777.408329] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 777.408329] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] yield resources [ 777.408329] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 777.408329] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] self.driver.spawn(context, instance, image_meta, [ 777.408329] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 777.408329] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 777.408329] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 777.408329] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] vm_ref = self.build_virtual_machine(instance, [ 777.408329] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 777.408869] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] vif_infos = vmwarevif.get_vif_info(self._session, [ 777.408869] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 777.408869] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] for vif in network_info: [ 777.408869] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 777.408869] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] return self._sync_wrapper(fn, *args, **kwargs) [ 777.408869] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 777.408869] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] self.wait() [ 777.408869] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 777.408869] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] self[:] = self._gt.wait() [ 777.408869] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 777.408869] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] return self._exit_event.wait() [ 777.408869] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 777.408869] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] current.throw(*self._exc) [ 777.409477] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 777.409477] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] result = function(*args, **kwargs) [ 777.409477] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 777.409477] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] return func(*args, **kwargs) [ 777.409477] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 777.409477] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] raise e [ 777.409477] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.409477] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] nwinfo = self.network_api.allocate_for_instance( [ 777.409477] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 777.409477] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] created_port_ids = self._update_ports_for_instance( [ 777.409477] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 777.409477] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] with excutils.save_and_reraise_exception(): [ 777.409477] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.410346] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] self.force_reraise() [ 777.410346] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.410346] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] raise self.value [ 777.410346] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 777.410346] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] updated_port = self._update_port( [ 777.410346] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.410346] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] _ensure_no_port_binding_failure(port) [ 777.410346] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.410346] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] raise exception.PortBindingFailed(port_id=port['id']) [ 777.410346] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] nova.exception.PortBindingFailed: Binding failed for port 52a0fa94-676e-4957-a12e-c9c1cd324b75, please check neutron logs for more information. [ 777.410346] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] [ 777.410346] env[62952]: INFO nova.compute.manager [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Terminating instance [ 777.411393] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Acquiring lock "refresh_cache-da96503a-8ae8-4b68-a782-631ee98754f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.411393] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Acquired lock "refresh_cache-da96503a-8ae8-4b68-a782-631ee98754f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.411393] env[62952]: DEBUG nova.network.neutron [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 777.659597] env[62952]: INFO nova.compute.manager [-] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Took 1.02 seconds to deallocate network for instance. [ 777.662158] env[62952]: DEBUG nova.compute.claims [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 777.662339] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.776974] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.777656] env[62952]: DEBUG nova.compute.manager [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 777.780558] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.754s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.824871] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99b746ec-d6b3-4724-ae66-095413be1bdd tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Lock "472e60f1-9b6c-415e-8a92-19235ad2e547" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.885s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.931678] env[62952]: DEBUG nova.network.neutron [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 778.019356] env[62952]: DEBUG nova.network.neutron [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.235290] env[62952]: INFO nova.compute.manager [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Rebuilding instance [ 778.274473] env[62952]: DEBUG nova.compute.manager [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 778.275408] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8cb321b-028e-4c54-a182-ed6d3b2ef709 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.286115] env[62952]: DEBUG nova.compute.utils [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 778.289230] env[62952]: DEBUG nova.compute.manager [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 778.289462] env[62952]: DEBUG nova.network.neutron [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 778.330047] env[62952]: DEBUG nova.compute.manager [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 778.333937] env[62952]: DEBUG nova.policy [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc539aa8b3b64dc99ba7491f18fa9949', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b23acb800ce4eb6ba431199fa360084', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 778.402159] env[62952]: DEBUG nova.compute.manager [req-4d27095a-968c-4ed2-9cc0-a48554ea22a0 req-864638c0-8611-4529-85df-8c6876c8199c service nova] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Received event network-changed-52a0fa94-676e-4957-a12e-c9c1cd324b75 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 778.402396] env[62952]: DEBUG nova.compute.manager [req-4d27095a-968c-4ed2-9cc0-a48554ea22a0 req-864638c0-8611-4529-85df-8c6876c8199c service nova] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Refreshing instance network info cache due to event network-changed-52a0fa94-676e-4957-a12e-c9c1cd324b75. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 778.402620] env[62952]: DEBUG oslo_concurrency.lockutils [req-4d27095a-968c-4ed2-9cc0-a48554ea22a0 req-864638c0-8611-4529-85df-8c6876c8199c service nova] Acquiring lock "refresh_cache-da96503a-8ae8-4b68-a782-631ee98754f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.522456] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Releasing lock "refresh_cache-da96503a-8ae8-4b68-a782-631ee98754f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.522853] env[62952]: DEBUG nova.compute.manager [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 778.523046] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 778.523356] env[62952]: DEBUG oslo_concurrency.lockutils [req-4d27095a-968c-4ed2-9cc0-a48554ea22a0 req-864638c0-8611-4529-85df-8c6876c8199c service nova] Acquired lock "refresh_cache-da96503a-8ae8-4b68-a782-631ee98754f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.523524] env[62952]: DEBUG nova.network.neutron [req-4d27095a-968c-4ed2-9cc0-a48554ea22a0 req-864638c0-8611-4529-85df-8c6876c8199c service nova] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Refreshing network info cache for port 52a0fa94-676e-4957-a12e-c9c1cd324b75 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 778.524677] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-21a793d6-56eb-4890-9dea-4d7e6a3bc847 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.536270] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a8061ff-71fe-404d-ba34-3bcf82835c3c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.561249] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance da96503a-8ae8-4b68-a782-631ee98754f7 could not be found. [ 778.561552] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 778.561740] env[62952]: INFO nova.compute.manager [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 778.561976] env[62952]: DEBUG oslo.service.loopingcall [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.564982] env[62952]: DEBUG nova.compute.manager [-] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 778.565102] env[62952]: DEBUG nova.network.neutron [-] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 778.587623] env[62952]: DEBUG nova.network.neutron [-] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 778.620533] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc964904-517c-4c16-91a1-3242eac87684 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.630099] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c4cc187-9416-4f06-95b0-e09105c3fc48 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.660718] env[62952]: DEBUG nova.network.neutron [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Successfully created port: 211aa71f-6364-48db-aeb2-5e51f549a03f {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 778.663202] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6201901f-7fe7-4a54-ba73-21931ad80bde {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.671027] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c45ed5d-1953-4694-8e34-7ca14c8c5908 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.685123] env[62952]: DEBUG nova.compute.provider_tree [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.790091] env[62952]: DEBUG nova.compute.manager [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 778.795393] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 778.795393] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f0dffb6-ff91-4fcf-9ca9-af998fc6ce9f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.802032] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 778.802032] env[62952]: value = "task-1263108" [ 778.802032] env[62952]: _type = "Task" [ 778.802032] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.810716] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263108, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.858610] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.046848] env[62952]: DEBUG nova.network.neutron [req-4d27095a-968c-4ed2-9cc0-a48554ea22a0 req-864638c0-8611-4529-85df-8c6876c8199c service nova] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 779.095386] env[62952]: DEBUG nova.network.neutron [-] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.162406] env[62952]: DEBUG nova.network.neutron [req-4d27095a-968c-4ed2-9cc0-a48554ea22a0 req-864638c0-8611-4529-85df-8c6876c8199c service nova] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.188014] env[62952]: DEBUG nova.scheduler.client.report [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.311542] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263108, 'name': PowerOffVM_Task, 'duration_secs': 0.123511} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.315022] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 779.315022] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 779.315022] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6122ff-78ae-4b8e-8cdd-b3412b0c2492 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.320801] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 779.321182] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c15a4a2e-44de-4961-9271-1524a8db55d6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.348666] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 779.348666] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 779.348666] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Deleting the datastore file [datastore1] 472e60f1-9b6c-415e-8a92-19235ad2e547 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 779.348666] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca4fe006-75ec-4b20-b037-e39cf5c99151 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.356248] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 779.356248] env[62952]: value = "task-1263110" [ 779.356248] env[62952]: _type = "Task" [ 779.356248] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.364714] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263110, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.581863] env[62952]: ERROR nova.compute.manager [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 211aa71f-6364-48db-aeb2-5e51f549a03f, please check neutron logs for more information. [ 779.581863] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 779.581863] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.581863] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 779.581863] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 779.581863] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 779.581863] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 779.581863] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 779.581863] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.581863] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 779.581863] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.581863] env[62952]: ERROR nova.compute.manager raise self.value [ 779.581863] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 779.581863] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 779.581863] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.581863] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 779.582367] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.582367] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 779.582367] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 211aa71f-6364-48db-aeb2-5e51f549a03f, please check neutron logs for more information. [ 779.582367] env[62952]: ERROR nova.compute.manager [ 779.582367] env[62952]: Traceback (most recent call last): [ 779.582367] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 779.582367] env[62952]: listener.cb(fileno) [ 779.582367] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.582367] env[62952]: result = function(*args, **kwargs) [ 779.582367] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 779.582367] env[62952]: return func(*args, **kwargs) [ 779.582367] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.582367] env[62952]: raise e [ 779.582367] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.582367] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 779.582367] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 779.582367] env[62952]: created_port_ids = self._update_ports_for_instance( [ 779.582367] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 779.582367] env[62952]: with excutils.save_and_reraise_exception(): [ 779.582367] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.582367] env[62952]: self.force_reraise() [ 779.582367] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.582367] env[62952]: raise self.value [ 779.582367] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 779.582367] env[62952]: updated_port = self._update_port( [ 779.582367] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.582367] env[62952]: _ensure_no_port_binding_failure(port) [ 779.582367] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.582367] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 779.583281] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 211aa71f-6364-48db-aeb2-5e51f549a03f, please check neutron logs for more information. [ 779.583281] env[62952]: Removing descriptor: 19 [ 779.597704] env[62952]: INFO nova.compute.manager [-] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Took 1.03 seconds to deallocate network for instance. [ 779.600183] env[62952]: DEBUG nova.compute.claims [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 779.600361] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.665530] env[62952]: DEBUG oslo_concurrency.lockutils [req-4d27095a-968c-4ed2-9cc0-a48554ea22a0 req-864638c0-8611-4529-85df-8c6876c8199c service nova] Releasing lock "refresh_cache-da96503a-8ae8-4b68-a782-631ee98754f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.665789] env[62952]: DEBUG nova.compute.manager [req-4d27095a-968c-4ed2-9cc0-a48554ea22a0 req-864638c0-8611-4529-85df-8c6876c8199c service nova] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Received event network-vif-deleted-52a0fa94-676e-4957-a12e-c9c1cd324b75 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 779.694016] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.913s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.694762] env[62952]: ERROR nova.compute.manager [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 00a9284f-3e43-4013-aa09-d38cfc6414ce, please check neutron logs for more information. [ 779.694762] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Traceback (most recent call last): [ 779.694762] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 779.694762] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] self.driver.spawn(context, instance, image_meta, [ 779.694762] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 779.694762] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] self._vmops.spawn(context, instance, image_meta, injected_files, [ 779.694762] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 779.694762] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] vm_ref = self.build_virtual_machine(instance, [ 779.694762] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 779.694762] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] vif_infos = vmwarevif.get_vif_info(self._session, [ 779.694762] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 779.695183] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] for vif in network_info: [ 779.695183] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 779.695183] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] return self._sync_wrapper(fn, *args, **kwargs) [ 779.695183] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 779.695183] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] self.wait() [ 779.695183] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 779.695183] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] self[:] = self._gt.wait() [ 779.695183] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 779.695183] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] return self._exit_event.wait() [ 779.695183] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 779.695183] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] result = hub.switch() [ 779.695183] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 779.695183] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] return self.greenlet.switch() [ 779.695596] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.695596] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] result = function(*args, **kwargs) [ 779.695596] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 779.695596] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] return func(*args, **kwargs) [ 779.695596] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.695596] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] raise e [ 779.695596] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.695596] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] nwinfo = self.network_api.allocate_for_instance( [ 779.695596] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 779.695596] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] created_port_ids = self._update_ports_for_instance( [ 779.695596] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 779.695596] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] with excutils.save_and_reraise_exception(): [ 779.695596] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.695997] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] self.force_reraise() [ 779.695997] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.695997] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] raise self.value [ 779.695997] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 779.695997] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] updated_port = self._update_port( [ 779.695997] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.695997] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] _ensure_no_port_binding_failure(port) [ 779.695997] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.695997] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] raise exception.PortBindingFailed(port_id=port['id']) [ 779.695997] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] nova.exception.PortBindingFailed: Binding failed for port 00a9284f-3e43-4013-aa09-d38cfc6414ce, please check neutron logs for more information. [ 779.695997] env[62952]: ERROR nova.compute.manager [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] [ 779.696479] env[62952]: DEBUG nova.compute.utils [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Binding failed for port 00a9284f-3e43-4013-aa09-d38cfc6414ce, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 779.697038] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.804s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.698688] env[62952]: INFO nova.compute.claims [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 779.701689] env[62952]: DEBUG nova.compute.manager [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Build of instance 5922d142-0b6f-4479-abcb-0c067a1ef837 was re-scheduled: Binding failed for port 00a9284f-3e43-4013-aa09-d38cfc6414ce, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 779.702187] env[62952]: DEBUG nova.compute.manager [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 779.702635] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Acquiring lock "refresh_cache-5922d142-0b6f-4479-abcb-0c067a1ef837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.702635] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Acquired lock "refresh_cache-5922d142-0b6f-4479-abcb-0c067a1ef837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.702844] env[62952]: DEBUG nova.network.neutron [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 779.800796] env[62952]: DEBUG nova.compute.manager [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 779.825362] env[62952]: DEBUG nova.virt.hardware [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 779.826017] env[62952]: DEBUG nova.virt.hardware [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 779.826017] env[62952]: DEBUG nova.virt.hardware [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 779.826017] env[62952]: DEBUG nova.virt.hardware [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 779.826166] env[62952]: DEBUG nova.virt.hardware [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 779.826198] env[62952]: DEBUG nova.virt.hardware [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 779.826400] env[62952]: DEBUG nova.virt.hardware [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 779.826556] env[62952]: DEBUG nova.virt.hardware [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 779.826717] env[62952]: DEBUG nova.virt.hardware [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 779.826872] env[62952]: DEBUG nova.virt.hardware [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 779.827050] env[62952]: DEBUG nova.virt.hardware [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 779.827895] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c7ed8c-28f2-4165-8e03-ef61846438d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.835748] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16d5a0ea-f60e-4d9b-8805-fc185eaba28d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.849066] env[62952]: ERROR nova.compute.manager [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 211aa71f-6364-48db-aeb2-5e51f549a03f, please check neutron logs for more information. [ 779.849066] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Traceback (most recent call last): [ 779.849066] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 779.849066] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] yield resources [ 779.849066] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 779.849066] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] self.driver.spawn(context, instance, image_meta, [ 779.849066] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 779.849066] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] self._vmops.spawn(context, instance, image_meta, injected_files, [ 779.849066] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 779.849066] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] vm_ref = self.build_virtual_machine(instance, [ 779.849066] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 779.849589] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] vif_infos = vmwarevif.get_vif_info(self._session, [ 779.849589] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 779.849589] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] for vif in network_info: [ 779.849589] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 779.849589] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] return self._sync_wrapper(fn, *args, **kwargs) [ 779.849589] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 779.849589] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] self.wait() [ 779.849589] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 779.849589] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] self[:] = self._gt.wait() [ 779.849589] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 779.849589] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] return self._exit_event.wait() [ 779.849589] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 779.849589] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] current.throw(*self._exc) [ 779.850061] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.850061] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] result = function(*args, **kwargs) [ 779.850061] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 779.850061] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] return func(*args, **kwargs) [ 779.850061] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.850061] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] raise e [ 779.850061] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.850061] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] nwinfo = self.network_api.allocate_for_instance( [ 779.850061] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 779.850061] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] created_port_ids = self._update_ports_for_instance( [ 779.850061] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 779.850061] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] with excutils.save_and_reraise_exception(): [ 779.850061] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.850459] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] self.force_reraise() [ 779.850459] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.850459] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] raise self.value [ 779.850459] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 779.850459] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] updated_port = self._update_port( [ 779.850459] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.850459] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] _ensure_no_port_binding_failure(port) [ 779.850459] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.850459] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] raise exception.PortBindingFailed(port_id=port['id']) [ 779.850459] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] nova.exception.PortBindingFailed: Binding failed for port 211aa71f-6364-48db-aeb2-5e51f549a03f, please check neutron logs for more information. [ 779.850459] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] [ 779.850459] env[62952]: INFO nova.compute.manager [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Terminating instance [ 779.851260] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Acquiring lock "refresh_cache-7f77fa00-27be-4a75-aac0-e1918f817837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.851414] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Acquired lock "refresh_cache-7f77fa00-27be-4a75-aac0-e1918f817837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.851576] env[62952]: DEBUG nova.network.neutron [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 779.863613] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263110, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09563} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.863833] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 779.864015] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 779.864190] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 780.221652] env[62952]: DEBUG nova.network.neutron [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 780.304979] env[62952]: DEBUG nova.network.neutron [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.368046] env[62952]: DEBUG nova.network.neutron [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 780.446058] env[62952]: DEBUG nova.network.neutron [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.454128] env[62952]: DEBUG nova.compute.manager [req-1db750a6-c612-4071-bb70-163e8c430a2b req-a6b6edb4-797c-4018-9d05-2ab1cb996129 service nova] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Received event network-changed-211aa71f-6364-48db-aeb2-5e51f549a03f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 780.454128] env[62952]: DEBUG nova.compute.manager [req-1db750a6-c612-4071-bb70-163e8c430a2b req-a6b6edb4-797c-4018-9d05-2ab1cb996129 service nova] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Refreshing instance network info cache due to event network-changed-211aa71f-6364-48db-aeb2-5e51f549a03f. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 780.454128] env[62952]: DEBUG oslo_concurrency.lockutils [req-1db750a6-c612-4071-bb70-163e8c430a2b req-a6b6edb4-797c-4018-9d05-2ab1cb996129 service nova] Acquiring lock "refresh_cache-7f77fa00-27be-4a75-aac0-e1918f817837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.806943] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Releasing lock "refresh_cache-5922d142-0b6f-4479-abcb-0c067a1ef837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.807256] env[62952]: DEBUG nova.compute.manager [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 780.807480] env[62952]: DEBUG nova.compute.manager [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 780.807683] env[62952]: DEBUG nova.network.neutron [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 780.822034] env[62952]: DEBUG nova.network.neutron [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 780.896019] env[62952]: DEBUG nova.virt.hardware [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 780.896019] env[62952]: DEBUG nova.virt.hardware [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 780.896019] env[62952]: DEBUG nova.virt.hardware [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 780.896019] env[62952]: DEBUG nova.virt.hardware [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 780.896276] env[62952]: DEBUG nova.virt.hardware [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 780.896276] env[62952]: DEBUG nova.virt.hardware [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 780.896463] env[62952]: DEBUG nova.virt.hardware [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 780.896619] env[62952]: DEBUG nova.virt.hardware [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 780.896777] env[62952]: DEBUG nova.virt.hardware [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 780.896930] env[62952]: DEBUG nova.virt.hardware [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 780.897109] env[62952]: DEBUG nova.virt.hardware [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 780.898249] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e1f02a-1d7f-4311-b425-21e3b256aac4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.907668] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b74db72-3140-45a8-aaca-02e3f76957e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.922724] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 780.928094] env[62952]: DEBUG oslo.service.loopingcall [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 780.930459] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 780.931196] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e1e9b3bb-c193-439c-96ab-57d329125f5b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.949883] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Releasing lock "refresh_cache-7f77fa00-27be-4a75-aac0-e1918f817837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.950367] env[62952]: DEBUG nova.compute.manager [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 780.950491] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 780.951285] env[62952]: DEBUG oslo_concurrency.lockutils [req-1db750a6-c612-4071-bb70-163e8c430a2b req-a6b6edb4-797c-4018-9d05-2ab1cb996129 service nova] Acquired lock "refresh_cache-7f77fa00-27be-4a75-aac0-e1918f817837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.951537] env[62952]: DEBUG nova.network.neutron [req-1db750a6-c612-4071-bb70-163e8c430a2b req-a6b6edb4-797c-4018-9d05-2ab1cb996129 service nova] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Refreshing network info cache for port 211aa71f-6364-48db-aeb2-5e51f549a03f {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 780.952461] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dc07873a-5d70-44d8-8924-957412013cf2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.956825] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 780.956825] env[62952]: value = "task-1263111" [ 780.956825] env[62952]: _type = "Task" [ 780.956825] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.966703] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49086f94-dcfc-4572-93a2-5e653f4efb1b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.980621] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263111, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.993492] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7f77fa00-27be-4a75-aac0-e1918f817837 could not be found. [ 780.993713] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 780.993891] env[62952]: INFO nova.compute.manager [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Took 0.04 seconds to destroy the instance on the hypervisor. [ 780.994151] env[62952]: DEBUG oslo.service.loopingcall [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 780.997039] env[62952]: DEBUG nova.compute.manager [-] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 780.997149] env[62952]: DEBUG nova.network.neutron [-] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 781.016755] env[62952]: DEBUG nova.network.neutron [-] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 781.044123] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7aa4370-73db-4742-80ad-7b3e01567083 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.051599] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f60a21f9-41bd-4f17-9b08-63aeaf971734 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.081881] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843bff92-f659-44c5-88fe-640a0ea2a2ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.088511] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18ab733-2937-40d7-a7e6-91d4ed67314b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.101014] env[62952]: DEBUG nova.compute.provider_tree [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 781.324515] env[62952]: DEBUG nova.network.neutron [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.466703] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263111, 'name': CreateVM_Task, 'duration_secs': 0.240459} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.466985] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 781.468721] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.468721] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.468721] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 781.468721] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b2cd547-ed56-41d4-8dac-dc3f1676604c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.472533] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 781.472533] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525fc00a-efd0-1633-5455-eb0ff5a72352" [ 781.472533] env[62952]: _type = "Task" [ 781.472533] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.475987] env[62952]: DEBUG nova.network.neutron [req-1db750a6-c612-4071-bb70-163e8c430a2b req-a6b6edb4-797c-4018-9d05-2ab1cb996129 service nova] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 781.482191] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525fc00a-efd0-1633-5455-eb0ff5a72352, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.519245] env[62952]: DEBUG nova.network.neutron [-] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.554904] env[62952]: DEBUG nova.network.neutron [req-1db750a6-c612-4071-bb70-163e8c430a2b req-a6b6edb4-797c-4018-9d05-2ab1cb996129 service nova] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.604309] env[62952]: DEBUG nova.scheduler.client.report [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 781.827342] env[62952]: INFO nova.compute.manager [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] [instance: 5922d142-0b6f-4479-abcb-0c067a1ef837] Took 1.02 seconds to deallocate network for instance. [ 781.983117] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525fc00a-efd0-1633-5455-eb0ff5a72352, 'name': SearchDatastore_Task, 'duration_secs': 0.008978} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.983421] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.983648] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 781.983871] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.984018] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.984190] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 781.984436] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d9770547-3bf0-4439-9bd9-34b2898ff344 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.992112] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 781.992289] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 781.992970] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61a3cbc1-b3bc-498b-b5bd-87bc420ad48e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.997604] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 781.997604] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52128b39-ad1e-499e-8545-a9bbcfe2a24c" [ 781.997604] env[62952]: _type = "Task" [ 781.997604] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.004480] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52128b39-ad1e-499e-8545-a9bbcfe2a24c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.022087] env[62952]: INFO nova.compute.manager [-] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Took 1.02 seconds to deallocate network for instance. [ 782.024149] env[62952]: DEBUG nova.compute.claims [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 782.024307] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.057234] env[62952]: DEBUG oslo_concurrency.lockutils [req-1db750a6-c612-4071-bb70-163e8c430a2b req-a6b6edb4-797c-4018-9d05-2ab1cb996129 service nova] Releasing lock "refresh_cache-7f77fa00-27be-4a75-aac0-e1918f817837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.057502] env[62952]: DEBUG nova.compute.manager [req-1db750a6-c612-4071-bb70-163e8c430a2b req-a6b6edb4-797c-4018-9d05-2ab1cb996129 service nova] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Received event network-vif-deleted-211aa71f-6364-48db-aeb2-5e51f549a03f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 782.109609] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.110116] env[62952]: DEBUG nova.compute.manager [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 782.112607] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.225s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.114091] env[62952]: INFO nova.compute.claims [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 782.509041] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52128b39-ad1e-499e-8545-a9bbcfe2a24c, 'name': SearchDatastore_Task, 'duration_secs': 0.007418} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.509838] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e90897b-6ce3-4f59-8e0e-e01e93cf42b7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.515236] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 782.515236] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52718f2e-3d55-b9c9-13a8-6a89e49d7a11" [ 782.515236] env[62952]: _type = "Task" [ 782.515236] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.523353] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52718f2e-3d55-b9c9-13a8-6a89e49d7a11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.620417] env[62952]: DEBUG nova.compute.utils [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 782.621860] env[62952]: DEBUG nova.compute.manager [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 782.622044] env[62952]: DEBUG nova.network.neutron [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 782.660774] env[62952]: DEBUG nova.policy [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e5b3cae6527f4f0db0794bbca1c5b6ac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ec2340d961b24ec9846d0326186bae6f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 782.860512] env[62952]: INFO nova.scheduler.client.report [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Deleted allocations for instance 5922d142-0b6f-4479-abcb-0c067a1ef837 [ 782.934436] env[62952]: DEBUG nova.network.neutron [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Successfully created port: 84777b2a-9e38-4048-b45c-f263fb473613 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 783.025338] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52718f2e-3d55-b9c9-13a8-6a89e49d7a11, 'name': SearchDatastore_Task, 'duration_secs': 0.008551} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.025608] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.025866] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 472e60f1-9b6c-415e-8a92-19235ad2e547/472e60f1-9b6c-415e-8a92-19235ad2e547.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 783.026136] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-26c290de-f45f-4c19-939d-7d4036a0439a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.032835] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 783.032835] env[62952]: value = "task-1263112" [ 783.032835] env[62952]: _type = "Task" [ 783.032835] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.040083] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263112, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.126057] env[62952]: DEBUG nova.compute.manager [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 783.372602] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc08d44e-cd69-4ad4-bd7e-5b2a57664180 tempest-ServerAddressesNegativeTestJSON-1808868233 tempest-ServerAddressesNegativeTestJSON-1808868233-project-member] Lock "5922d142-0b6f-4479-abcb-0c067a1ef837" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.464s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.483897] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9d1d39-f0c3-46b1-b8e5-f76aa7aa865f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.490683] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aac9460-2eac-4fcf-b305-4a1228a84cf9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.525297] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-481f0658-b018-4721-bb3d-d348042b94aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.533647] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e8838e-1f75-4f8b-a669-d837b32acaa5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.544433] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263112, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446025} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.551919] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 472e60f1-9b6c-415e-8a92-19235ad2e547/472e60f1-9b6c-415e-8a92-19235ad2e547.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 783.552187] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 783.553220] env[62952]: DEBUG nova.compute.provider_tree [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 783.553820] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e94cb393-91e4-4f8f-8284-bb35f84514e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.560900] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 783.560900] env[62952]: value = "task-1263113" [ 783.560900] env[62952]: _type = "Task" [ 783.560900] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.569658] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263113, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.733132] env[62952]: DEBUG nova.compute.manager [req-0510691b-a871-4bc5-89a4-bd070a2c55cb req-51d5f342-b920-4451-aa91-5d1912ec66b2 service nova] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Received event network-changed-84777b2a-9e38-4048-b45c-f263fb473613 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 783.733132] env[62952]: DEBUG nova.compute.manager [req-0510691b-a871-4bc5-89a4-bd070a2c55cb req-51d5f342-b920-4451-aa91-5d1912ec66b2 service nova] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Refreshing instance network info cache due to event network-changed-84777b2a-9e38-4048-b45c-f263fb473613. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 783.733999] env[62952]: DEBUG oslo_concurrency.lockutils [req-0510691b-a871-4bc5-89a4-bd070a2c55cb req-51d5f342-b920-4451-aa91-5d1912ec66b2 service nova] Acquiring lock "refresh_cache-879dc564-0a5e-48aa-92ff-b02435445a68" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.734483] env[62952]: DEBUG oslo_concurrency.lockutils [req-0510691b-a871-4bc5-89a4-bd070a2c55cb req-51d5f342-b920-4451-aa91-5d1912ec66b2 service nova] Acquired lock "refresh_cache-879dc564-0a5e-48aa-92ff-b02435445a68" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.735630] env[62952]: DEBUG nova.network.neutron [req-0510691b-a871-4bc5-89a4-bd070a2c55cb req-51d5f342-b920-4451-aa91-5d1912ec66b2 service nova] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Refreshing network info cache for port 84777b2a-9e38-4048-b45c-f263fb473613 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 783.864697] env[62952]: ERROR nova.compute.manager [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 84777b2a-9e38-4048-b45c-f263fb473613, please check neutron logs for more information. [ 783.864697] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 783.864697] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 783.864697] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 783.864697] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 783.864697] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 783.864697] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 783.864697] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 783.864697] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.864697] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 783.864697] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.864697] env[62952]: ERROR nova.compute.manager raise self.value [ 783.864697] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 783.864697] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 783.864697] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.864697] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 783.865143] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.865143] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 783.865143] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 84777b2a-9e38-4048-b45c-f263fb473613, please check neutron logs for more information. [ 783.865143] env[62952]: ERROR nova.compute.manager [ 783.865143] env[62952]: Traceback (most recent call last): [ 783.865143] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 783.865143] env[62952]: listener.cb(fileno) [ 783.865143] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 783.865143] env[62952]: result = function(*args, **kwargs) [ 783.865143] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 783.865143] env[62952]: return func(*args, **kwargs) [ 783.865143] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 783.865143] env[62952]: raise e [ 783.865143] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 783.865143] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 783.865143] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 783.865143] env[62952]: created_port_ids = self._update_ports_for_instance( [ 783.865143] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 783.865143] env[62952]: with excutils.save_and_reraise_exception(): [ 783.865143] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.865143] env[62952]: self.force_reraise() [ 783.865143] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.865143] env[62952]: raise self.value [ 783.865143] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 783.865143] env[62952]: updated_port = self._update_port( [ 783.865143] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.865143] env[62952]: _ensure_no_port_binding_failure(port) [ 783.865143] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.865143] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 783.866046] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 84777b2a-9e38-4048-b45c-f263fb473613, please check neutron logs for more information. [ 783.866046] env[62952]: Removing descriptor: 19 [ 783.883019] env[62952]: DEBUG nova.compute.manager [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 784.057694] env[62952]: DEBUG nova.scheduler.client.report [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 784.071285] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263113, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078651} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.071548] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 784.072379] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aac02cd-ffea-4fa5-8862-423e668d0cc7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.092820] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Reconfiguring VM instance instance-00000024 to attach disk [datastore1] 472e60f1-9b6c-415e-8a92-19235ad2e547/472e60f1-9b6c-415e-8a92-19235ad2e547.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 784.093262] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac5d9416-c6a8-4781-bc9d-2ebd1d427cfc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.112698] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 784.112698] env[62952]: value = "task-1263114" [ 784.112698] env[62952]: _type = "Task" [ 784.112698] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.121148] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263114, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.135708] env[62952]: DEBUG nova.compute.manager [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 784.165509] env[62952]: DEBUG nova.virt.hardware [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 784.165757] env[62952]: DEBUG nova.virt.hardware [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 784.165910] env[62952]: DEBUG nova.virt.hardware [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 784.166099] env[62952]: DEBUG nova.virt.hardware [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 784.166251] env[62952]: DEBUG nova.virt.hardware [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 784.166388] env[62952]: DEBUG nova.virt.hardware [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 784.166587] env[62952]: DEBUG nova.virt.hardware [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 784.166739] env[62952]: DEBUG nova.virt.hardware [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 784.166897] env[62952]: DEBUG nova.virt.hardware [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 784.167061] env[62952]: DEBUG nova.virt.hardware [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 784.167230] env[62952]: DEBUG nova.virt.hardware [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 784.168578] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1a5b08f-4060-4668-b2ab-ead5503824da {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.176799] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e9e18d2-92eb-4d62-8f76-75508b7e933b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.190217] env[62952]: ERROR nova.compute.manager [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 84777b2a-9e38-4048-b45c-f263fb473613, please check neutron logs for more information. [ 784.190217] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Traceback (most recent call last): [ 784.190217] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 784.190217] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] yield resources [ 784.190217] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 784.190217] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] self.driver.spawn(context, instance, image_meta, [ 784.190217] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 784.190217] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] self._vmops.spawn(context, instance, image_meta, injected_files, [ 784.190217] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 784.190217] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] vm_ref = self.build_virtual_machine(instance, [ 784.190217] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 784.190538] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] vif_infos = vmwarevif.get_vif_info(self._session, [ 784.190538] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 784.190538] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] for vif in network_info: [ 784.190538] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 784.190538] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] return self._sync_wrapper(fn, *args, **kwargs) [ 784.190538] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 784.190538] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] self.wait() [ 784.190538] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 784.190538] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] self[:] = self._gt.wait() [ 784.190538] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 784.190538] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] return self._exit_event.wait() [ 784.190538] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 784.190538] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] current.throw(*self._exc) [ 784.190855] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 784.190855] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] result = function(*args, **kwargs) [ 784.190855] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 784.190855] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] return func(*args, **kwargs) [ 784.190855] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 784.190855] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] raise e [ 784.190855] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 784.190855] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] nwinfo = self.network_api.allocate_for_instance( [ 784.190855] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 784.190855] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] created_port_ids = self._update_ports_for_instance( [ 784.190855] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 784.190855] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] with excutils.save_and_reraise_exception(): [ 784.190855] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 784.191179] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] self.force_reraise() [ 784.191179] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 784.191179] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] raise self.value [ 784.191179] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 784.191179] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] updated_port = self._update_port( [ 784.191179] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 784.191179] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] _ensure_no_port_binding_failure(port) [ 784.191179] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 784.191179] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] raise exception.PortBindingFailed(port_id=port['id']) [ 784.191179] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] nova.exception.PortBindingFailed: Binding failed for port 84777b2a-9e38-4048-b45c-f263fb473613, please check neutron logs for more information. [ 784.191179] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] [ 784.191179] env[62952]: INFO nova.compute.manager [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Terminating instance [ 784.192482] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Acquiring lock "refresh_cache-879dc564-0a5e-48aa-92ff-b02435445a68" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.258075] env[62952]: DEBUG nova.network.neutron [req-0510691b-a871-4bc5-89a4-bd070a2c55cb req-51d5f342-b920-4451-aa91-5d1912ec66b2 service nova] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 784.391406] env[62952]: DEBUG nova.network.neutron [req-0510691b-a871-4bc5-89a4-bd070a2c55cb req-51d5f342-b920-4451-aa91-5d1912ec66b2 service nova] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.404970] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.563016] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.563453] env[62952]: DEBUG nova.compute.manager [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 784.566225] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.882s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.626366] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263114, 'name': ReconfigVM_Task, 'duration_secs': 0.268925} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.626652] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Reconfigured VM instance instance-00000024 to attach disk [datastore1] 472e60f1-9b6c-415e-8a92-19235ad2e547/472e60f1-9b6c-415e-8a92-19235ad2e547.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 784.627491] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f62a5d96-9471-43d6-b7d6-dd70b5d3e428 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.633375] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 784.633375] env[62952]: value = "task-1263115" [ 784.633375] env[62952]: _type = "Task" [ 784.633375] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.641700] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263115, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.895552] env[62952]: DEBUG oslo_concurrency.lockutils [req-0510691b-a871-4bc5-89a4-bd070a2c55cb req-51d5f342-b920-4451-aa91-5d1912ec66b2 service nova] Releasing lock "refresh_cache-879dc564-0a5e-48aa-92ff-b02435445a68" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.896276] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Acquired lock "refresh_cache-879dc564-0a5e-48aa-92ff-b02435445a68" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.896276] env[62952]: DEBUG nova.network.neutron [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 785.072599] env[62952]: DEBUG nova.compute.utils [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 785.072599] env[62952]: DEBUG nova.compute.manager [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 785.072599] env[62952]: DEBUG nova.network.neutron [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 785.143428] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263115, 'name': Rename_Task, 'duration_secs': 0.125516} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.143817] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 785.144312] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-49a30c14-3a88-4358-8121-e8fa5a3dcb36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.151739] env[62952]: DEBUG nova.policy [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3acdbe29682040eeb84db57e295646d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a657bc8e636f44e8abbfc487a7ed296d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 785.154167] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 785.154167] env[62952]: value = "task-1263116" [ 785.154167] env[62952]: _type = "Task" [ 785.154167] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.163782] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263116, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.423411] env[62952]: DEBUG nova.network.neutron [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 785.448550] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90375d4b-d0bf-43a8-bd04-a072577a3654 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.456358] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16b7521a-6ecc-492c-94eb-ea61122a2f79 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.487360] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3490e32-c3c2-4424-a1fb-bf8084d7a2f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.494613] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47babec9-a63f-4f17-b5dd-0717494431fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.507735] env[62952]: DEBUG nova.compute.provider_tree [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.576440] env[62952]: DEBUG nova.network.neutron [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.579613] env[62952]: DEBUG nova.compute.manager [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 785.629693] env[62952]: DEBUG nova.network.neutron [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Successfully created port: aa429812-d328-4234-92d8-01fbd7263815 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 785.664500] env[62952]: DEBUG oslo_vmware.api [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263116, 'name': PowerOnVM_Task, 'duration_secs': 0.446786} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.665043] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 785.665545] env[62952]: DEBUG nova.compute.manager [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 785.666438] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47553865-83b5-4abf-9ef4-9502259818d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.797699] env[62952]: DEBUG nova.compute.manager [req-3850142f-ddae-4dfe-ac20-0f5ba4466994 req-8922ccf6-831f-4748-8d15-997b632dda47 service nova] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Received event network-vif-deleted-84777b2a-9e38-4048-b45c-f263fb473613 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 786.014085] env[62952]: DEBUG nova.scheduler.client.report [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 786.083106] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Releasing lock "refresh_cache-879dc564-0a5e-48aa-92ff-b02435445a68" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.083106] env[62952]: DEBUG nova.compute.manager [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 786.083106] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 786.083106] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2f572afa-e280-4758-bb6d-dd5a28c517a0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.087145] env[62952]: INFO nova.virt.block_device [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Booting with volume e824dcab-a7d4-4de6-a654-9f5fba715ab6 at /dev/sda [ 786.116453] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af276a37-4c85-4eed-94b4-5c774f2ab911 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.150132] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 879dc564-0a5e-48aa-92ff-b02435445a68 could not be found. [ 786.156103] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 786.156618] env[62952]: INFO nova.compute.manager [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Took 0.07 seconds to destroy the instance on the hypervisor. [ 786.156905] env[62952]: DEBUG oslo.service.loopingcall [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 786.157206] env[62952]: DEBUG nova.compute.manager [-] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 786.157310] env[62952]: DEBUG nova.network.neutron [-] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 786.183201] env[62952]: DEBUG nova.network.neutron [-] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 786.187139] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.207583] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a01d58ad-496f-4f1b-9732-d49e6207f24a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.217918] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d770a3-7c1b-4bfb-8d66-76563028023a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.241638] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a7ae6658-d5b9-40bd-b90e-3f0f1cdc21e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.249845] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d57ac24-7a57-447c-9d53-b33b824e222f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.272158] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-584ff5a5-eb9b-42f7-9b47-5021badb2cf6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.278216] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-808ed6cf-026b-4db1-8c14-9370cd537ed5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.292799] env[62952]: DEBUG nova.virt.block_device [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Updating existing volume attachment record: 6da8b9e1-c3a2-47be-8125-32a3d764136f {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 786.519520] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.953s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.520182] env[62952]: ERROR nova.compute.manager [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 672ecb52-afb2-4f75-8227-ca1b0e7c21a1, please check neutron logs for more information. [ 786.520182] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Traceback (most recent call last): [ 786.520182] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 786.520182] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] self.driver.spawn(context, instance, image_meta, [ 786.520182] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 786.520182] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 786.520182] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 786.520182] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] vm_ref = self.build_virtual_machine(instance, [ 786.520182] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 786.520182] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] vif_infos = vmwarevif.get_vif_info(self._session, [ 786.520182] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 786.520546] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] for vif in network_info: [ 786.520546] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 786.520546] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] return self._sync_wrapper(fn, *args, **kwargs) [ 786.520546] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 786.520546] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] self.wait() [ 786.520546] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 786.520546] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] self[:] = self._gt.wait() [ 786.520546] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 786.520546] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] return self._exit_event.wait() [ 786.520546] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 786.520546] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] current.throw(*self._exc) [ 786.520546] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 786.520546] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] result = function(*args, **kwargs) [ 786.520948] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 786.520948] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] return func(*args, **kwargs) [ 786.520948] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 786.520948] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] raise e [ 786.520948] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 786.520948] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] nwinfo = self.network_api.allocate_for_instance( [ 786.520948] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 786.520948] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] created_port_ids = self._update_ports_for_instance( [ 786.520948] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 786.520948] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] with excutils.save_and_reraise_exception(): [ 786.520948] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 786.520948] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] self.force_reraise() [ 786.520948] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 786.521392] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] raise self.value [ 786.521392] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 786.521392] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] updated_port = self._update_port( [ 786.521392] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 786.521392] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] _ensure_no_port_binding_failure(port) [ 786.521392] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 786.521392] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] raise exception.PortBindingFailed(port_id=port['id']) [ 786.521392] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] nova.exception.PortBindingFailed: Binding failed for port 672ecb52-afb2-4f75-8227-ca1b0e7c21a1, please check neutron logs for more information. [ 786.521392] env[62952]: ERROR nova.compute.manager [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] [ 786.521392] env[62952]: DEBUG nova.compute.utils [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Binding failed for port 672ecb52-afb2-4f75-8227-ca1b0e7c21a1, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 786.522278] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.173s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.524138] env[62952]: INFO nova.compute.claims [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 786.528434] env[62952]: DEBUG nova.compute.manager [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Build of instance a458a658-96f3-41b2-a9ee-2a6dd52bfeb4 was re-scheduled: Binding failed for port 672ecb52-afb2-4f75-8227-ca1b0e7c21a1, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 786.528434] env[62952]: DEBUG nova.compute.manager [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 786.528434] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Acquiring lock "refresh_cache-a458a658-96f3-41b2-a9ee-2a6dd52bfeb4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.528434] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Acquired lock "refresh_cache-a458a658-96f3-41b2-a9ee-2a6dd52bfeb4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.528720] env[62952]: DEBUG nova.network.neutron [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 786.599050] env[62952]: ERROR nova.compute.manager [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port aa429812-d328-4234-92d8-01fbd7263815, please check neutron logs for more information. [ 786.599050] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 786.599050] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 786.599050] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 786.599050] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 786.599050] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 786.599050] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 786.599050] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 786.599050] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 786.599050] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 786.599050] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 786.599050] env[62952]: ERROR nova.compute.manager raise self.value [ 786.599050] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 786.599050] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 786.599050] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 786.599050] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 786.599742] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 786.599742] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 786.599742] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port aa429812-d328-4234-92d8-01fbd7263815, please check neutron logs for more information. [ 786.599742] env[62952]: ERROR nova.compute.manager [ 786.599742] env[62952]: Traceback (most recent call last): [ 786.599742] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 786.599742] env[62952]: listener.cb(fileno) [ 786.599742] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 786.599742] env[62952]: result = function(*args, **kwargs) [ 786.599742] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 786.599742] env[62952]: return func(*args, **kwargs) [ 786.599742] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 786.599742] env[62952]: raise e [ 786.599742] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 786.599742] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 786.599742] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 786.599742] env[62952]: created_port_ids = self._update_ports_for_instance( [ 786.599742] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 786.599742] env[62952]: with excutils.save_and_reraise_exception(): [ 786.599742] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 786.599742] env[62952]: self.force_reraise() [ 786.599742] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 786.599742] env[62952]: raise self.value [ 786.599742] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 786.599742] env[62952]: updated_port = self._update_port( [ 786.599742] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 786.599742] env[62952]: _ensure_no_port_binding_failure(port) [ 786.599742] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 786.599742] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 786.600669] env[62952]: nova.exception.PortBindingFailed: Binding failed for port aa429812-d328-4234-92d8-01fbd7263815, please check neutron logs for more information. [ 786.600669] env[62952]: Removing descriptor: 19 [ 786.688161] env[62952]: DEBUG nova.network.neutron [-] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.048893] env[62952]: DEBUG nova.network.neutron [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 787.125571] env[62952]: DEBUG nova.network.neutron [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.191322] env[62952]: INFO nova.compute.manager [-] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Took 1.03 seconds to deallocate network for instance. [ 787.193700] env[62952]: DEBUG nova.compute.claims [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 787.195017] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.430497] env[62952]: INFO nova.compute.manager [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Rebuilding instance [ 787.470634] env[62952]: DEBUG nova.compute.manager [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 787.471508] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea936916-33e0-4e3f-a0a7-5d91df160abe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.628232] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Releasing lock "refresh_cache-a458a658-96f3-41b2-a9ee-2a6dd52bfeb4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.628861] env[62952]: DEBUG nova.compute.manager [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 787.628861] env[62952]: DEBUG nova.compute.manager [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 787.628861] env[62952]: DEBUG nova.network.neutron [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 787.643762] env[62952]: DEBUG nova.network.neutron [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 787.824690] env[62952]: DEBUG nova.compute.manager [req-596e8b4a-1e67-427b-9e00-c3ebb88fad73 req-2036acbd-8136-4d50-afa2-cffd00eff1a7 service nova] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Received event network-changed-aa429812-d328-4234-92d8-01fbd7263815 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 787.824875] env[62952]: DEBUG nova.compute.manager [req-596e8b4a-1e67-427b-9e00-c3ebb88fad73 req-2036acbd-8136-4d50-afa2-cffd00eff1a7 service nova] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Refreshing instance network info cache due to event network-changed-aa429812-d328-4234-92d8-01fbd7263815. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 787.825036] env[62952]: DEBUG oslo_concurrency.lockutils [req-596e8b4a-1e67-427b-9e00-c3ebb88fad73 req-2036acbd-8136-4d50-afa2-cffd00eff1a7 service nova] Acquiring lock "refresh_cache-3822c92d-560b-4d61-92e9-7c694bc7a33a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.825181] env[62952]: DEBUG oslo_concurrency.lockutils [req-596e8b4a-1e67-427b-9e00-c3ebb88fad73 req-2036acbd-8136-4d50-afa2-cffd00eff1a7 service nova] Acquired lock "refresh_cache-3822c92d-560b-4d61-92e9-7c694bc7a33a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.825333] env[62952]: DEBUG nova.network.neutron [req-596e8b4a-1e67-427b-9e00-c3ebb88fad73 req-2036acbd-8136-4d50-afa2-cffd00eff1a7 service nova] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Refreshing network info cache for port aa429812-d328-4234-92d8-01fbd7263815 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 787.843270] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58fe63b0-7bbb-4624-b5fc-b038691adc07 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.852302] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ed120b-e9bf-4c13-8d91-3d5109d7ffdd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.883874] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f4ece78-1ed7-4f34-ac71-6f3f6c2177ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.891325] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d7cccf-cb9d-4d6d-b4af-22589e2a49eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.904451] env[62952]: DEBUG nova.compute.provider_tree [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.982323] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 787.982615] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-82c03d94-f88c-4fb5-882d-eeef660ce291 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.990228] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Waiting for the task: (returnval){ [ 787.990228] env[62952]: value = "task-1263117" [ 787.990228] env[62952]: _type = "Task" [ 787.990228] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.000138] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': task-1263117, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.152857] env[62952]: DEBUG nova.network.neutron [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.346532] env[62952]: DEBUG nova.network.neutron [req-596e8b4a-1e67-427b-9e00-c3ebb88fad73 req-2036acbd-8136-4d50-afa2-cffd00eff1a7 service nova] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 788.409125] env[62952]: DEBUG nova.scheduler.client.report [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 788.413325] env[62952]: DEBUG nova.compute.manager [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 788.413977] env[62952]: DEBUG nova.virt.hardware [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 788.414365] env[62952]: DEBUG nova.virt.hardware [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 788.414641] env[62952]: DEBUG nova.virt.hardware [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 788.414930] env[62952]: DEBUG nova.virt.hardware [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 788.415196] env[62952]: DEBUG nova.virt.hardware [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 788.415445] env[62952]: DEBUG nova.virt.hardware [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 788.415771] env[62952]: DEBUG nova.virt.hardware [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 788.416050] env[62952]: DEBUG nova.virt.hardware [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 788.416325] env[62952]: DEBUG nova.virt.hardware [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 788.416597] env[62952]: DEBUG nova.virt.hardware [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 788.416919] env[62952]: DEBUG nova.virt.hardware [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 788.418058] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d5d7533-df1b-4bd2-89d1-d98a6d4573bc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.430025] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4389409b-070e-4fd8-acd4-4d0b3175dd5b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.442889] env[62952]: ERROR nova.compute.manager [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port aa429812-d328-4234-92d8-01fbd7263815, please check neutron logs for more information. [ 788.442889] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Traceback (most recent call last): [ 788.442889] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 788.442889] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] yield resources [ 788.442889] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 788.442889] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] self.driver.spawn(context, instance, image_meta, [ 788.442889] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 788.442889] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 788.442889] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 788.442889] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] vm_ref = self.build_virtual_machine(instance, [ 788.442889] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 788.443337] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] vif_infos = vmwarevif.get_vif_info(self._session, [ 788.443337] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 788.443337] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] for vif in network_info: [ 788.443337] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 788.443337] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] return self._sync_wrapper(fn, *args, **kwargs) [ 788.443337] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 788.443337] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] self.wait() [ 788.443337] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 788.443337] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] self[:] = self._gt.wait() [ 788.443337] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 788.443337] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] return self._exit_event.wait() [ 788.443337] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 788.443337] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] current.throw(*self._exc) [ 788.443732] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 788.443732] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] result = function(*args, **kwargs) [ 788.443732] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 788.443732] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] return func(*args, **kwargs) [ 788.443732] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 788.443732] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] raise e [ 788.443732] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 788.443732] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] nwinfo = self.network_api.allocate_for_instance( [ 788.443732] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 788.443732] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] created_port_ids = self._update_ports_for_instance( [ 788.443732] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 788.443732] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] with excutils.save_and_reraise_exception(): [ 788.443732] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 788.444134] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] self.force_reraise() [ 788.444134] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 788.444134] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] raise self.value [ 788.444134] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 788.444134] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] updated_port = self._update_port( [ 788.444134] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 788.444134] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] _ensure_no_port_binding_failure(port) [ 788.444134] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 788.444134] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] raise exception.PortBindingFailed(port_id=port['id']) [ 788.444134] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] nova.exception.PortBindingFailed: Binding failed for port aa429812-d328-4234-92d8-01fbd7263815, please check neutron logs for more information. [ 788.444134] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] [ 788.444134] env[62952]: INFO nova.compute.manager [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Terminating instance [ 788.444944] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Acquiring lock "refresh_cache-3822c92d-560b-4d61-92e9-7c694bc7a33a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.472340] env[62952]: DEBUG nova.network.neutron [req-596e8b4a-1e67-427b-9e00-c3ebb88fad73 req-2036acbd-8136-4d50-afa2-cffd00eff1a7 service nova] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.499766] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': task-1263117, 'name': PowerOffVM_Task, 'duration_secs': 0.234588} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.501632] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 788.501632] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 788.501632] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef63e99-83f4-4f32-b8e2-a84079c36e6f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.507960] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 788.508209] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8e6f7ed1-80c6-49ad-b3ec-4a9073080e61 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.530499] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 788.530703] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 788.530884] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Deleting the datastore file [datastore1] 472e60f1-9b6c-415e-8a92-19235ad2e547 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 788.531150] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fe9946e3-04c3-495a-921a-59619f3652a0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.537449] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Waiting for the task: (returnval){ [ 788.537449] env[62952]: value = "task-1263119" [ 788.537449] env[62952]: _type = "Task" [ 788.537449] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.546702] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': task-1263119, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.655723] env[62952]: INFO nova.compute.manager [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] [instance: a458a658-96f3-41b2-a9ee-2a6dd52bfeb4] Took 1.03 seconds to deallocate network for instance. [ 788.920037] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.920037] env[62952]: DEBUG nova.compute.manager [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 788.922286] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.811s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.977610] env[62952]: DEBUG oslo_concurrency.lockutils [req-596e8b4a-1e67-427b-9e00-c3ebb88fad73 req-2036acbd-8136-4d50-afa2-cffd00eff1a7 service nova] Releasing lock "refresh_cache-3822c92d-560b-4d61-92e9-7c694bc7a33a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.977891] env[62952]: DEBUG nova.compute.manager [req-596e8b4a-1e67-427b-9e00-c3ebb88fad73 req-2036acbd-8136-4d50-afa2-cffd00eff1a7 service nova] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Received event network-vif-deleted-aa429812-d328-4234-92d8-01fbd7263815 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 788.978538] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Acquired lock "refresh_cache-3822c92d-560b-4d61-92e9-7c694bc7a33a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.978739] env[62952]: DEBUG nova.network.neutron [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 789.049081] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': task-1263119, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095392} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.049081] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 789.049081] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 789.049081] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 789.429222] env[62952]: DEBUG nova.compute.utils [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 789.436017] env[62952]: DEBUG nova.compute.manager [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 789.436017] env[62952]: DEBUG nova.network.neutron [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 789.486309] env[62952]: DEBUG nova.policy [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ff3a53dbede4951b062ac5df0d1c011', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '02ecd052199f4ec4a291755df3cd4d2e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 789.503452] env[62952]: DEBUG nova.network.neutron [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 789.621610] env[62952]: DEBUG nova.network.neutron [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.687355] env[62952]: INFO nova.scheduler.client.report [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Deleted allocations for instance a458a658-96f3-41b2-a9ee-2a6dd52bfeb4 [ 789.786620] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b88089c-7cfa-42bc-86a8-7696cef52b4a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.795490] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290aa1df-1aed-49c7-98b2-843b1f39a745 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.834536] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e78713-7b99-4e50-bff7-28dba28c3297 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.839873] env[62952]: DEBUG nova.network.neutron [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Successfully created port: 57491018-5fb5-4504-a04f-c8729d155e31 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 789.845654] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1548def9-af47-4dd3-b8cc-79502e4554fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.860348] env[62952]: DEBUG nova.compute.provider_tree [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.865403] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Acquiring lock "67b729be-fb66-4ed1-bbea-e62216d460d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.865687] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Lock "67b729be-fb66-4ed1-bbea-e62216d460d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.935288] env[62952]: DEBUG nova.compute.manager [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 790.093635] env[62952]: DEBUG nova.virt.hardware [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 790.093890] env[62952]: DEBUG nova.virt.hardware [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 790.094061] env[62952]: DEBUG nova.virt.hardware [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.094246] env[62952]: DEBUG nova.virt.hardware [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 790.094392] env[62952]: DEBUG nova.virt.hardware [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.094534] env[62952]: DEBUG nova.virt.hardware [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 790.094737] env[62952]: DEBUG nova.virt.hardware [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 790.094892] env[62952]: DEBUG nova.virt.hardware [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 790.095471] env[62952]: DEBUG nova.virt.hardware [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 790.095689] env[62952]: DEBUG nova.virt.hardware [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 790.095871] env[62952]: DEBUG nova.virt.hardware [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.096757] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970ba4ee-5a6d-422a-abbd-231ea82c2615 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.105138] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6f9b17e-3e39-4e32-becb-8f60d5f08a26 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.118318] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 790.124057] env[62952]: DEBUG oslo.service.loopingcall [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.124597] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 790.125053] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Releasing lock "refresh_cache-3822c92d-560b-4d61-92e9-7c694bc7a33a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.125592] env[62952]: DEBUG nova.compute.manager [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 790.126069] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-61b9c457-0448-441f-927f-dc37ec16831a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.137671] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c521aab0-6d85-405a-987e-32b51dd417bc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.147329] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e574e3fd-7c64-42ca-a71c-02eb54bd2de1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.158700] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 790.158700] env[62952]: value = "task-1263120" [ 790.158700] env[62952]: _type = "Task" [ 790.158700] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.166577] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263120, 'name': CreateVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.170142] env[62952]: WARNING nova.virt.vmwareapi.driver [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 3822c92d-560b-4d61-92e9-7c694bc7a33a could not be found. [ 790.170347] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 790.170601] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-76174b4a-e55d-44c6-9e32-9aae1f450a0d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.177817] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cda7ff5-55bf-469f-9897-efd78adab18e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.199720] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3822c92d-560b-4d61-92e9-7c694bc7a33a could not be found. [ 790.199947] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 790.200141] env[62952]: INFO nova.compute.manager [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Took 0.07 seconds to destroy the instance on the hypervisor. [ 790.200408] env[62952]: DEBUG oslo.service.loopingcall [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.200598] env[62952]: DEBUG nova.compute.manager [-] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 790.200690] env[62952]: DEBUG nova.network.neutron [-] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 790.202576] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3708b3e-d10e-46ec-81a5-c951d904aade tempest-AttachInterfacesUnderV243Test-1252657424 tempest-AttachInterfacesUnderV243Test-1252657424-project-member] Lock "a458a658-96f3-41b2-a9ee-2a6dd52bfeb4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.611s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.222433] env[62952]: DEBUG nova.network.neutron [-] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 790.366647] env[62952]: DEBUG nova.scheduler.client.report [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 790.652096] env[62952]: DEBUG nova.compute.manager [req-e7ce8201-e407-4a68-bb57-6f107000fb19 req-8f72aef7-ac4c-43be-908e-db1cd5d76600 service nova] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Received event network-changed-57491018-5fb5-4504-a04f-c8729d155e31 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 790.652286] env[62952]: DEBUG nova.compute.manager [req-e7ce8201-e407-4a68-bb57-6f107000fb19 req-8f72aef7-ac4c-43be-908e-db1cd5d76600 service nova] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Refreshing instance network info cache due to event network-changed-57491018-5fb5-4504-a04f-c8729d155e31. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 790.652502] env[62952]: DEBUG oslo_concurrency.lockutils [req-e7ce8201-e407-4a68-bb57-6f107000fb19 req-8f72aef7-ac4c-43be-908e-db1cd5d76600 service nova] Acquiring lock "refresh_cache-9979dd1d-9e99-47cf-a81a-fd7d89531f51" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.652643] env[62952]: DEBUG oslo_concurrency.lockutils [req-e7ce8201-e407-4a68-bb57-6f107000fb19 req-8f72aef7-ac4c-43be-908e-db1cd5d76600 service nova] Acquired lock "refresh_cache-9979dd1d-9e99-47cf-a81a-fd7d89531f51" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.652800] env[62952]: DEBUG nova.network.neutron [req-e7ce8201-e407-4a68-bb57-6f107000fb19 req-8f72aef7-ac4c-43be-908e-db1cd5d76600 service nova] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Refreshing network info cache for port 57491018-5fb5-4504-a04f-c8729d155e31 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 790.668813] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263120, 'name': CreateVM_Task, 'duration_secs': 0.229184} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.668978] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 790.669414] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.669572] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.669938] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 790.670195] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47963484-76f6-420c-8b69-4e03a271b413 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.675128] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Waiting for the task: (returnval){ [ 790.675128] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]527f9c8c-73bb-3fb5-ff82-921b9f99c497" [ 790.675128] env[62952]: _type = "Task" [ 790.675128] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.683679] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]527f9c8c-73bb-3fb5-ff82-921b9f99c497, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.705378] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 790.726017] env[62952]: DEBUG nova.network.neutron [-] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.758148] env[62952]: ERROR nova.compute.manager [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 57491018-5fb5-4504-a04f-c8729d155e31, please check neutron logs for more information. [ 790.758148] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 790.758148] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.758148] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 790.758148] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 790.758148] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 790.758148] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 790.758148] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 790.758148] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.758148] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 790.758148] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.758148] env[62952]: ERROR nova.compute.manager raise self.value [ 790.758148] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 790.758148] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 790.758148] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.758148] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 790.758805] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.758805] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 790.758805] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 57491018-5fb5-4504-a04f-c8729d155e31, please check neutron logs for more information. [ 790.758805] env[62952]: ERROR nova.compute.manager [ 790.758805] env[62952]: Traceback (most recent call last): [ 790.758805] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 790.758805] env[62952]: listener.cb(fileno) [ 790.758805] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 790.758805] env[62952]: result = function(*args, **kwargs) [ 790.758805] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 790.758805] env[62952]: return func(*args, **kwargs) [ 790.758805] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 790.758805] env[62952]: raise e [ 790.758805] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.758805] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 790.758805] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 790.758805] env[62952]: created_port_ids = self._update_ports_for_instance( [ 790.758805] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 790.758805] env[62952]: with excutils.save_and_reraise_exception(): [ 790.758805] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.758805] env[62952]: self.force_reraise() [ 790.758805] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.758805] env[62952]: raise self.value [ 790.758805] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 790.758805] env[62952]: updated_port = self._update_port( [ 790.758805] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.758805] env[62952]: _ensure_no_port_binding_failure(port) [ 790.758805] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.758805] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 790.759768] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 57491018-5fb5-4504-a04f-c8729d155e31, please check neutron logs for more information. [ 790.759768] env[62952]: Removing descriptor: 19 [ 790.871831] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.949s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.872488] env[62952]: ERROR nova.compute.manager [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b3154985-2dd0-4f1e-b7fe-d36d43a66c89, please check neutron logs for more information. [ 790.872488] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Traceback (most recent call last): [ 790.872488] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 790.872488] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] self.driver.spawn(context, instance, image_meta, [ 790.872488] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 790.872488] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 790.872488] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 790.872488] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] vm_ref = self.build_virtual_machine(instance, [ 790.872488] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 790.872488] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] vif_infos = vmwarevif.get_vif_info(self._session, [ 790.872488] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 790.872799] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] for vif in network_info: [ 790.872799] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 790.872799] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] return self._sync_wrapper(fn, *args, **kwargs) [ 790.872799] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 790.872799] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] self.wait() [ 790.872799] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 790.872799] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] self[:] = self._gt.wait() [ 790.872799] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 790.872799] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] return self._exit_event.wait() [ 790.872799] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 790.872799] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] current.throw(*self._exc) [ 790.872799] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 790.872799] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] result = function(*args, **kwargs) [ 790.873144] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 790.873144] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] return func(*args, **kwargs) [ 790.873144] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 790.873144] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] raise e [ 790.873144] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.873144] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] nwinfo = self.network_api.allocate_for_instance( [ 790.873144] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 790.873144] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] created_port_ids = self._update_ports_for_instance( [ 790.873144] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 790.873144] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] with excutils.save_and_reraise_exception(): [ 790.873144] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.873144] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] self.force_reraise() [ 790.873144] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.873466] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] raise self.value [ 790.873466] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 790.873466] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] updated_port = self._update_port( [ 790.873466] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.873466] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] _ensure_no_port_binding_failure(port) [ 790.873466] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.873466] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] raise exception.PortBindingFailed(port_id=port['id']) [ 790.873466] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] nova.exception.PortBindingFailed: Binding failed for port b3154985-2dd0-4f1e-b7fe-d36d43a66c89, please check neutron logs for more information. [ 790.873466] env[62952]: ERROR nova.compute.manager [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] [ 790.873466] env[62952]: DEBUG nova.compute.utils [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Binding failed for port b3154985-2dd0-4f1e-b7fe-d36d43a66c89, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 790.874773] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.212s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.877932] env[62952]: DEBUG nova.compute.manager [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Build of instance dd88d70d-17a3-4059-9092-5d6cec8b1ae4 was re-scheduled: Binding failed for port b3154985-2dd0-4f1e-b7fe-d36d43a66c89, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 790.879073] env[62952]: DEBUG nova.compute.manager [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 790.879316] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquiring lock "refresh_cache-dd88d70d-17a3-4059-9092-5d6cec8b1ae4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.879463] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquired lock "refresh_cache-dd88d70d-17a3-4059-9092-5d6cec8b1ae4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.879689] env[62952]: DEBUG nova.network.neutron [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 790.946027] env[62952]: DEBUG nova.compute.manager [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 790.973821] env[62952]: DEBUG nova.virt.hardware [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:32:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='f9152906-6db9-472a-ac84-7e1b6a1c26a6',id=33,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-233560411',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 790.974083] env[62952]: DEBUG nova.virt.hardware [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 790.974223] env[62952]: DEBUG nova.virt.hardware [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.974395] env[62952]: DEBUG nova.virt.hardware [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 790.974541] env[62952]: DEBUG nova.virt.hardware [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.974683] env[62952]: DEBUG nova.virt.hardware [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 790.974892] env[62952]: DEBUG nova.virt.hardware [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 790.975170] env[62952]: DEBUG nova.virt.hardware [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 790.975356] env[62952]: DEBUG nova.virt.hardware [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 790.975519] env[62952]: DEBUG nova.virt.hardware [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 790.975688] env[62952]: DEBUG nova.virt.hardware [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.976846] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d0f2b1e-3fec-4bd7-aee8-65719f9f4e78 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.984815] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e00ac33-0bb7-46f8-8f4f-9d59f1fecd43 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.999773] env[62952]: ERROR nova.compute.manager [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 57491018-5fb5-4504-a04f-c8729d155e31, please check neutron logs for more information. [ 790.999773] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Traceback (most recent call last): [ 790.999773] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 790.999773] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] yield resources [ 790.999773] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 790.999773] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] self.driver.spawn(context, instance, image_meta, [ 790.999773] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 790.999773] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] self._vmops.spawn(context, instance, image_meta, injected_files, [ 790.999773] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 790.999773] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] vm_ref = self.build_virtual_machine(instance, [ 790.999773] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 791.000155] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] vif_infos = vmwarevif.get_vif_info(self._session, [ 791.000155] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 791.000155] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] for vif in network_info: [ 791.000155] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 791.000155] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] return self._sync_wrapper(fn, *args, **kwargs) [ 791.000155] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 791.000155] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] self.wait() [ 791.000155] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 791.000155] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] self[:] = self._gt.wait() [ 791.000155] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 791.000155] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] return self._exit_event.wait() [ 791.000155] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 791.000155] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] current.throw(*self._exc) [ 791.000603] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 791.000603] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] result = function(*args, **kwargs) [ 791.000603] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 791.000603] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] return func(*args, **kwargs) [ 791.000603] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 791.000603] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] raise e [ 791.000603] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.000603] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] nwinfo = self.network_api.allocate_for_instance( [ 791.000603] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 791.000603] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] created_port_ids = self._update_ports_for_instance( [ 791.000603] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 791.000603] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] with excutils.save_and_reraise_exception(): [ 791.000603] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.000992] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] self.force_reraise() [ 791.000992] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.000992] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] raise self.value [ 791.000992] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 791.000992] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] updated_port = self._update_port( [ 791.000992] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.000992] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] _ensure_no_port_binding_failure(port) [ 791.000992] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.000992] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] raise exception.PortBindingFailed(port_id=port['id']) [ 791.000992] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] nova.exception.PortBindingFailed: Binding failed for port 57491018-5fb5-4504-a04f-c8729d155e31, please check neutron logs for more information. [ 791.000992] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] [ 791.000992] env[62952]: INFO nova.compute.manager [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Terminating instance [ 791.005987] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Acquiring lock "refresh_cache-9979dd1d-9e99-47cf-a81a-fd7d89531f51" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.175777] env[62952]: DEBUG nova.network.neutron [req-e7ce8201-e407-4a68-bb57-6f107000fb19 req-8f72aef7-ac4c-43be-908e-db1cd5d76600 service nova] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 791.186483] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]527f9c8c-73bb-3fb5-ff82-921b9f99c497, 'name': SearchDatastore_Task, 'duration_secs': 0.008567} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.186940] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.187245] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 791.187540] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.187744] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.188016] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 791.188332] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a869d4e1-aa71-4227-b2eb-2d41edbea5d1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.196277] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 791.196900] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 791.197278] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ddd8f31-87df-4a62-a5f8-0e25839eb9f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.202307] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Waiting for the task: (returnval){ [ 791.202307] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523ad685-bac2-71a9-b7aa-5729e8e4dffe" [ 791.202307] env[62952]: _type = "Task" [ 791.202307] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.214493] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523ad685-bac2-71a9-b7aa-5729e8e4dffe, 'name': SearchDatastore_Task, 'duration_secs': 0.007883} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.217166] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d29c672e-beb0-4f7c-8f18-755702f766b3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.222197] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Waiting for the task: (returnval){ [ 791.222197] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52eaed99-8dc6-3cea-1742-038d0fc4916b" [ 791.222197] env[62952]: _type = "Task" [ 791.222197] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.230900] env[62952]: INFO nova.compute.manager [-] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Took 1.03 seconds to deallocate network for instance. [ 791.230900] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52eaed99-8dc6-3cea-1742-038d0fc4916b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.233169] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.278345] env[62952]: DEBUG nova.network.neutron [req-e7ce8201-e407-4a68-bb57-6f107000fb19 req-8f72aef7-ac4c-43be-908e-db1cd5d76600 service nova] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.410895] env[62952]: DEBUG nova.network.neutron [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 791.527179] env[62952]: DEBUG nova.network.neutron [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.661969] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6fa1e2-7c5b-4afd-af99-43939772b85b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.669202] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d143a4-fd7a-4793-914d-115c9443f053 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.701044] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed72315-eed1-4615-a36a-4f91e310cb77 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.710129] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a166f64-3133-4864-9a9a-b49228575fc0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.725663] env[62952]: DEBUG nova.compute.provider_tree [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.736163] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52eaed99-8dc6-3cea-1742-038d0fc4916b, 'name': SearchDatastore_Task, 'duration_secs': 0.007987} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.736163] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.736163] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 472e60f1-9b6c-415e-8a92-19235ad2e547/472e60f1-9b6c-415e-8a92-19235ad2e547.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 791.736163] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5a577aa1-427c-4162-89bd-d96b8850be54 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.744032] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Waiting for the task: (returnval){ [ 791.744032] env[62952]: value = "task-1263121" [ 791.744032] env[62952]: _type = "Task" [ 791.744032] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.752392] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': task-1263121, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.781080] env[62952]: DEBUG oslo_concurrency.lockutils [req-e7ce8201-e407-4a68-bb57-6f107000fb19 req-8f72aef7-ac4c-43be-908e-db1cd5d76600 service nova] Releasing lock "refresh_cache-9979dd1d-9e99-47cf-a81a-fd7d89531f51" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.781976] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Acquired lock "refresh_cache-9979dd1d-9e99-47cf-a81a-fd7d89531f51" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.781976] env[62952]: DEBUG nova.network.neutron [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 791.803278] env[62952]: INFO nova.compute.manager [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Took 0.57 seconds to detach 1 volumes for instance. [ 791.805407] env[62952]: DEBUG nova.compute.claims [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 791.805610] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.031023] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Releasing lock "refresh_cache-dd88d70d-17a3-4059-9092-5d6cec8b1ae4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.031023] env[62952]: DEBUG nova.compute.manager [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 792.031239] env[62952]: DEBUG nova.compute.manager [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 792.031303] env[62952]: DEBUG nova.network.neutron [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 792.053440] env[62952]: DEBUG nova.network.neutron [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 792.234021] env[62952]: DEBUG nova.scheduler.client.report [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 792.254880] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': task-1263121, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.437855} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.254880] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 472e60f1-9b6c-415e-8a92-19235ad2e547/472e60f1-9b6c-415e-8a92-19235ad2e547.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 792.254880] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 792.256124] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e5d618d1-ecc5-48a2-9a08-9ab19779d7a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.262364] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Waiting for the task: (returnval){ [ 792.262364] env[62952]: value = "task-1263122" [ 792.262364] env[62952]: _type = "Task" [ 792.262364] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.270171] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': task-1263122, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.300442] env[62952]: DEBUG nova.network.neutron [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 792.351433] env[62952]: DEBUG nova.network.neutron [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.555700] env[62952]: DEBUG nova.network.neutron [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.679452] env[62952]: DEBUG nova.compute.manager [req-7d1a3cc6-28f9-40a1-8c81-9fde19309d07 req-364b4dee-37d0-4197-81de-e95e10887e08 service nova] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Received event network-vif-deleted-57491018-5fb5-4504-a04f-c8729d155e31 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 792.739555] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.865s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.740250] env[62952]: ERROR nova.compute.manager [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0613de0a-f8a0-4c51-a045-3442c42f143a, please check neutron logs for more information. [ 792.740250] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Traceback (most recent call last): [ 792.740250] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 792.740250] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] self.driver.spawn(context, instance, image_meta, [ 792.740250] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 792.740250] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 792.740250] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 792.740250] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] vm_ref = self.build_virtual_machine(instance, [ 792.740250] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 792.740250] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 792.740250] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 792.740620] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] for vif in network_info: [ 792.740620] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 792.740620] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] return self._sync_wrapper(fn, *args, **kwargs) [ 792.740620] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 792.740620] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] self.wait() [ 792.740620] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 792.740620] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] self[:] = self._gt.wait() [ 792.740620] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 792.740620] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] return self._exit_event.wait() [ 792.740620] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 792.740620] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] current.throw(*self._exc) [ 792.740620] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 792.740620] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] result = function(*args, **kwargs) [ 792.741355] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 792.741355] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] return func(*args, **kwargs) [ 792.741355] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 792.741355] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] raise e [ 792.741355] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 792.741355] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] nwinfo = self.network_api.allocate_for_instance( [ 792.741355] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 792.741355] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] created_port_ids = self._update_ports_for_instance( [ 792.741355] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 792.741355] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] with excutils.save_and_reraise_exception(): [ 792.741355] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.741355] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] self.force_reraise() [ 792.741355] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.741896] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] raise self.value [ 792.741896] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 792.741896] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] updated_port = self._update_port( [ 792.741896] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.741896] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] _ensure_no_port_binding_failure(port) [ 792.741896] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.741896] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] raise exception.PortBindingFailed(port_id=port['id']) [ 792.741896] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] nova.exception.PortBindingFailed: Binding failed for port 0613de0a-f8a0-4c51-a045-3442c42f143a, please check neutron logs for more information. [ 792.741896] env[62952]: ERROR nova.compute.manager [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] [ 792.741896] env[62952]: DEBUG nova.compute.utils [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Binding failed for port 0613de0a-f8a0-4c51-a045-3442c42f143a, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 792.742652] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.884s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.744172] env[62952]: INFO nova.compute.claims [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 792.747164] env[62952]: DEBUG nova.compute.manager [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Build of instance b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6 was re-scheduled: Binding failed for port 0613de0a-f8a0-4c51-a045-3442c42f143a, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 792.747677] env[62952]: DEBUG nova.compute.manager [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 792.747869] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "refresh_cache-b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.747971] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "refresh_cache-b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.748136] env[62952]: DEBUG nova.network.neutron [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 792.773552] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': task-1263122, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066815} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.773686] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 792.774759] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6de040-cf7b-4dc9-a938-8733cdff1359 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.796248] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Reconfiguring VM instance instance-00000024 to attach disk [datastore2] 472e60f1-9b6c-415e-8a92-19235ad2e547/472e60f1-9b6c-415e-8a92-19235ad2e547.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.797142] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2dbadcd6-d946-471b-94e7-0f6a8ad69dcd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.817610] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Waiting for the task: (returnval){ [ 792.817610] env[62952]: value = "task-1263123" [ 792.817610] env[62952]: _type = "Task" [ 792.817610] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.827503] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': task-1263123, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.855223] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Releasing lock "refresh_cache-9979dd1d-9e99-47cf-a81a-fd7d89531f51" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.855787] env[62952]: DEBUG nova.compute.manager [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 792.855906] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 792.856213] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ada84fc2-2948-409f-a31e-e6a78db977dc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.864551] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25cf0659-88b7-4525-b38a-0f1e37559ab3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.887112] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9979dd1d-9e99-47cf-a81a-fd7d89531f51 could not be found. [ 792.887348] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 792.887530] env[62952]: INFO nova.compute.manager [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Took 0.03 seconds to destroy the instance on the hypervisor. [ 792.887772] env[62952]: DEBUG oslo.service.loopingcall [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.887986] env[62952]: DEBUG nova.compute.manager [-] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 792.888097] env[62952]: DEBUG nova.network.neutron [-] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 792.905552] env[62952]: DEBUG nova.network.neutron [-] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 793.059105] env[62952]: INFO nova.compute.manager [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: dd88d70d-17a3-4059-9092-5d6cec8b1ae4] Took 1.03 seconds to deallocate network for instance. [ 793.270380] env[62952]: DEBUG nova.network.neutron [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 793.331437] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': task-1263123, 'name': ReconfigVM_Task, 'duration_secs': 0.266469} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.332101] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Reconfigured VM instance instance-00000024 to attach disk [datastore2] 472e60f1-9b6c-415e-8a92-19235ad2e547/472e60f1-9b6c-415e-8a92-19235ad2e547.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.332345] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09ec6893-dfb3-415d-95b2-628fd592f5ae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.338630] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Waiting for the task: (returnval){ [ 793.338630] env[62952]: value = "task-1263124" [ 793.338630] env[62952]: _type = "Task" [ 793.338630] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.346429] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': task-1263124, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.360834] env[62952]: DEBUG nova.network.neutron [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.409317] env[62952]: DEBUG nova.network.neutron [-] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.850202] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': task-1263124, 'name': Rename_Task, 'duration_secs': 0.146053} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.850478] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 793.850808] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5351744-14b5-4324-a974-29b91ad65a94 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.859054] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Waiting for the task: (returnval){ [ 793.859054] env[62952]: value = "task-1263125" [ 793.859054] env[62952]: _type = "Task" [ 793.859054] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.862698] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "refresh_cache-b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.862911] env[62952]: DEBUG nova.compute.manager [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 793.863116] env[62952]: DEBUG nova.compute.manager [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 793.863297] env[62952]: DEBUG nova.network.neutron [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 793.872883] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': task-1263125, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.883851] env[62952]: DEBUG nova.network.neutron [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 793.912769] env[62952]: INFO nova.compute.manager [-] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Took 1.02 seconds to deallocate network for instance. [ 793.915122] env[62952]: DEBUG nova.compute.claims [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 793.915307] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.087100] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42612d6-5fc0-411c-b16e-426a4d2e851f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.091132] env[62952]: INFO nova.scheduler.client.report [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Deleted allocations for instance dd88d70d-17a3-4059-9092-5d6cec8b1ae4 [ 794.101081] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4fb3ad6-8a27-4d88-b348-d110ca2ee8ac {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.133021] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f13f330-d273-45b8-b218-188b68d88f23 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.141691] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbfc48de-648d-43e0-b034-0bde8910c961 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.156333] env[62952]: DEBUG nova.compute.provider_tree [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 794.369842] env[62952]: DEBUG oslo_vmware.api [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Task: {'id': task-1263125, 'name': PowerOnVM_Task, 'duration_secs': 0.457632} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.370328] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 794.370695] env[62952]: DEBUG nova.compute.manager [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 794.371606] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbafd092-3653-4483-a53f-dd677c29e569 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.386730] env[62952]: DEBUG nova.network.neutron [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.598814] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e200c9-5245-427e-b82d-2242d0c2dc32 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Lock "dd88d70d-17a3-4059-9092-5d6cec8b1ae4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.137s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.684699] env[62952]: ERROR nova.scheduler.client.report [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [req-b7451f16-ee4f-4012-b2bc-56874ec18e81] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 615127b5-dc82-4491-9c8f-4e6d0caa0690. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b7451f16-ee4f-4012-b2bc-56874ec18e81"}]} [ 794.706276] env[62952]: DEBUG nova.scheduler.client.report [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Refreshing inventories for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 794.721112] env[62952]: DEBUG nova.scheduler.client.report [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Updating ProviderTree inventory for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 794.721760] env[62952]: DEBUG nova.compute.provider_tree [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 794.741477] env[62952]: DEBUG nova.scheduler.client.report [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Refreshing aggregate associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 794.762552] env[62952]: DEBUG nova.scheduler.client.report [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Refreshing trait associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 794.888698] env[62952]: INFO nova.compute.manager [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6] Took 1.03 seconds to deallocate network for instance. [ 794.895123] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.102160] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 795.138076] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77557c82-95a9-4dd4-ac38-cc8569af34c1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.146627] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-915452bb-9482-4e52-bd4d-09b7d77a0b08 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.177134] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b4e9d4-9a3d-4baf-8f67-fdd4d4d5a281 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.184485] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55931c9-b9b8-4035-b8c9-a78d216046b1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.199343] env[62952]: DEBUG nova.compute.provider_tree [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 795.456024] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquiring lock "472e60f1-9b6c-415e-8a92-19235ad2e547" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.456024] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Lock "472e60f1-9b6c-415e-8a92-19235ad2e547" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.456153] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquiring lock "472e60f1-9b6c-415e-8a92-19235ad2e547-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.456779] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Lock "472e60f1-9b6c-415e-8a92-19235ad2e547-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.456779] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Lock "472e60f1-9b6c-415e-8a92-19235ad2e547-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.460024] env[62952]: INFO nova.compute.manager [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Terminating instance [ 795.462699] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquiring lock "refresh_cache-472e60f1-9b6c-415e-8a92-19235ad2e547" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.462699] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquired lock "refresh_cache-472e60f1-9b6c-415e-8a92-19235ad2e547" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.462699] env[62952]: DEBUG nova.network.neutron [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 795.498336] env[62952]: DEBUG oslo_concurrency.lockutils [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquiring lock "2b11ad9b-63d8-40da-8f68-3d8260c63bc7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.499246] env[62952]: DEBUG oslo_concurrency.lockutils [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Lock "2b11ad9b-63d8-40da-8f68-3d8260c63bc7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.625450] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.724115] env[62952]: ERROR nova.scheduler.client.report [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [req-053fa60d-2585-4a4a-b9ae-4572a61b0e00] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 615127b5-dc82-4491-9c8f-4e6d0caa0690. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-053fa60d-2585-4a4a-b9ae-4572a61b0e00"}]} [ 795.743321] env[62952]: DEBUG nova.scheduler.client.report [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Refreshing inventories for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 795.759809] env[62952]: DEBUG nova.scheduler.client.report [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Updating ProviderTree inventory for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 795.760054] env[62952]: DEBUG nova.compute.provider_tree [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 795.772193] env[62952]: DEBUG nova.scheduler.client.report [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Refreshing aggregate associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 795.792725] env[62952]: DEBUG nova.scheduler.client.report [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Refreshing trait associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 795.932832] env[62952]: INFO nova.scheduler.client.report [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleted allocations for instance b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6 [ 795.988264] env[62952]: DEBUG nova.network.neutron [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 796.078373] env[62952]: DEBUG nova.network.neutron [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.150300] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbbe688b-ef75-4f34-be25-298c9b8bfa66 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.159508] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de812f4-7a9c-4063-ac27-fa13f29cf496 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.190303] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6e4e4f-3e3b-43f1-9120-67e8808bf018 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.200023] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0858812-0259-4b1d-8727-2f2eed0a9b4f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.211640] env[62952]: DEBUG nova.compute.provider_tree [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 796.445896] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7a54149e-a8b4-4c65-b680-e72812f317c8 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "b7a5eeff-39b3-4e9a-956b-bcbd4492c3f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.556s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.585323] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Releasing lock "refresh_cache-472e60f1-9b6c-415e-8a92-19235ad2e547" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.585556] env[62952]: DEBUG nova.compute.manager [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 796.585741] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 796.586628] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458b116c-58fe-44ab-9a8a-5ade36fe2e84 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.594855] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 796.595109] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f9828df-1517-4b54-8fbc-112a2cd5d45e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.601145] env[62952]: DEBUG oslo_vmware.api [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 796.601145] env[62952]: value = "task-1263126" [ 796.601145] env[62952]: _type = "Task" [ 796.601145] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.610318] env[62952]: DEBUG oslo_vmware.api [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263126, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.760669] env[62952]: DEBUG nova.scheduler.client.report [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Updated inventory for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with generation 68 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 796.761027] env[62952]: DEBUG nova.compute.provider_tree [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Updating resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 generation from 68 to 69 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 796.761143] env[62952]: DEBUG nova.compute.provider_tree [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 796.952197] env[62952]: DEBUG nova.compute.manager [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 797.112929] env[62952]: DEBUG oslo_vmware.api [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263126, 'name': PowerOffVM_Task, 'duration_secs': 0.10942} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.113758] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 797.113939] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 797.115675] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-30304797-f3b4-47a9-bac0-0779c01705da {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.142828] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 797.142895] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 797.143078] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Deleting the datastore file [datastore2] 472e60f1-9b6c-415e-8a92-19235ad2e547 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 797.143335] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6068f018-f6b0-45a1-9115-8548d9cb54d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.151318] env[62952]: DEBUG oslo_vmware.api [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for the task: (returnval){ [ 797.151318] env[62952]: value = "task-1263128" [ 797.151318] env[62952]: _type = "Task" [ 797.151318] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.161159] env[62952]: DEBUG oslo_vmware.api [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263128, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.267589] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.525s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.268143] env[62952]: DEBUG nova.compute.manager [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 797.272081] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.672s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.470792] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.610898] env[62952]: DEBUG oslo_concurrency.lockutils [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.611157] env[62952]: DEBUG oslo_concurrency.lockutils [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.662677] env[62952]: DEBUG oslo_vmware.api [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Task: {'id': task-1263128, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127227} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.666537] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 797.666833] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 797.667020] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 797.667219] env[62952]: INFO nova.compute.manager [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Took 1.08 seconds to destroy the instance on the hypervisor. [ 797.667463] env[62952]: DEBUG oslo.service.loopingcall [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.667653] env[62952]: DEBUG nova.compute.manager [-] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 797.667762] env[62952]: DEBUG nova.network.neutron [-] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 797.688020] env[62952]: DEBUG nova.network.neutron [-] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 797.780134] env[62952]: DEBUG nova.compute.utils [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 797.781677] env[62952]: DEBUG nova.compute.manager [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 797.781935] env[62952]: DEBUG nova.network.neutron [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 797.822627] env[62952]: DEBUG nova.policy [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f47d1bc5e40145efaf2f4f9f5d7d100f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd0cac7c3e85d44989f8c7ddb80acfae6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 798.075286] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8914fba-4297-4e9c-9003-c894b160d6e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.083225] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e76853-19ab-4b98-855f-7aa99969ac45 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.116419] env[62952]: DEBUG nova.network.neutron [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Successfully created port: 6629a780-8685-4991-9015-0601155c4978 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 798.118832] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-198cf8c2-e8c2-4d39-8d81-66b0920563e0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.127341] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb99c04-63a8-4df7-8756-d2eed42bcd7f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.143354] env[62952]: DEBUG nova.compute.provider_tree [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.190102] env[62952]: DEBUG nova.network.neutron [-] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.284839] env[62952]: DEBUG nova.compute.manager [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 798.612097] env[62952]: DEBUG oslo_concurrency.lockutils [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "dd700ee9-0957-4210-be50-e4c8ac3ca456" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.612346] env[62952]: DEBUG oslo_concurrency.lockutils [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "dd700ee9-0957-4210-be50-e4c8ac3ca456" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.646695] env[62952]: DEBUG nova.scheduler.client.report [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 798.692957] env[62952]: INFO nova.compute.manager [-] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Took 1.03 seconds to deallocate network for instance. [ 798.799532] env[62952]: DEBUG nova.compute.manager [req-7da18bda-da8c-44fc-ae18-e43293a9ef7d req-4749ba80-e9d7-4a3f-aac4-d4a85f22d5ea service nova] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Received event network-changed-6629a780-8685-4991-9015-0601155c4978 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 798.802154] env[62952]: DEBUG nova.compute.manager [req-7da18bda-da8c-44fc-ae18-e43293a9ef7d req-4749ba80-e9d7-4a3f-aac4-d4a85f22d5ea service nova] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Refreshing instance network info cache due to event network-changed-6629a780-8685-4991-9015-0601155c4978. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 798.802154] env[62952]: DEBUG oslo_concurrency.lockutils [req-7da18bda-da8c-44fc-ae18-e43293a9ef7d req-4749ba80-e9d7-4a3f-aac4-d4a85f22d5ea service nova] Acquiring lock "refresh_cache-ebff0c87-8e58-4a76-a350-43a3d660debd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.802154] env[62952]: DEBUG oslo_concurrency.lockutils [req-7da18bda-da8c-44fc-ae18-e43293a9ef7d req-4749ba80-e9d7-4a3f-aac4-d4a85f22d5ea service nova] Acquired lock "refresh_cache-ebff0c87-8e58-4a76-a350-43a3d660debd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.802154] env[62952]: DEBUG nova.network.neutron [req-7da18bda-da8c-44fc-ae18-e43293a9ef7d req-4749ba80-e9d7-4a3f-aac4-d4a85f22d5ea service nova] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Refreshing network info cache for port 6629a780-8685-4991-9015-0601155c4978 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 798.975627] env[62952]: ERROR nova.compute.manager [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6629a780-8685-4991-9015-0601155c4978, please check neutron logs for more information. [ 798.975627] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 798.975627] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 798.975627] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 798.975627] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 798.975627] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 798.975627] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 798.975627] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 798.975627] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 798.975627] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 798.975627] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 798.975627] env[62952]: ERROR nova.compute.manager raise self.value [ 798.975627] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 798.975627] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 798.975627] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 798.975627] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 798.976168] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 798.976168] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 798.976168] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6629a780-8685-4991-9015-0601155c4978, please check neutron logs for more information. [ 798.976168] env[62952]: ERROR nova.compute.manager [ 798.976168] env[62952]: Traceback (most recent call last): [ 798.976168] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 798.976168] env[62952]: listener.cb(fileno) [ 798.976168] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 798.976168] env[62952]: result = function(*args, **kwargs) [ 798.976168] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 798.976168] env[62952]: return func(*args, **kwargs) [ 798.976168] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 798.976168] env[62952]: raise e [ 798.976168] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 798.976168] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 798.976168] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 798.976168] env[62952]: created_port_ids = self._update_ports_for_instance( [ 798.976168] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 798.976168] env[62952]: with excutils.save_and_reraise_exception(): [ 798.976168] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 798.976168] env[62952]: self.force_reraise() [ 798.976168] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 798.976168] env[62952]: raise self.value [ 798.976168] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 798.976168] env[62952]: updated_port = self._update_port( [ 798.976168] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 798.976168] env[62952]: _ensure_no_port_binding_failure(port) [ 798.976168] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 798.976168] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 798.977164] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 6629a780-8685-4991-9015-0601155c4978, please check neutron logs for more information. [ 798.977164] env[62952]: Removing descriptor: 19 [ 799.152848] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.881s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.153489] env[62952]: ERROR nova.compute.manager [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 52a0fa94-676e-4957-a12e-c9c1cd324b75, please check neutron logs for more information. [ 799.153489] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Traceback (most recent call last): [ 799.153489] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 799.153489] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] self.driver.spawn(context, instance, image_meta, [ 799.153489] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 799.153489] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 799.153489] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 799.153489] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] vm_ref = self.build_virtual_machine(instance, [ 799.153489] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 799.153489] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] vif_infos = vmwarevif.get_vif_info(self._session, [ 799.153489] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 799.153926] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] for vif in network_info: [ 799.153926] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 799.153926] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] return self._sync_wrapper(fn, *args, **kwargs) [ 799.153926] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 799.153926] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] self.wait() [ 799.153926] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 799.153926] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] self[:] = self._gt.wait() [ 799.153926] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 799.153926] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] return self._exit_event.wait() [ 799.153926] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 799.153926] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] current.throw(*self._exc) [ 799.153926] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 799.153926] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] result = function(*args, **kwargs) [ 799.154349] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 799.154349] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] return func(*args, **kwargs) [ 799.154349] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 799.154349] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] raise e [ 799.154349] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 799.154349] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] nwinfo = self.network_api.allocate_for_instance( [ 799.154349] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 799.154349] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] created_port_ids = self._update_ports_for_instance( [ 799.154349] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 799.154349] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] with excutils.save_and_reraise_exception(): [ 799.154349] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 799.154349] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] self.force_reraise() [ 799.154349] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 799.154757] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] raise self.value [ 799.154757] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 799.154757] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] updated_port = self._update_port( [ 799.154757] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 799.154757] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] _ensure_no_port_binding_failure(port) [ 799.154757] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 799.154757] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] raise exception.PortBindingFailed(port_id=port['id']) [ 799.154757] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] nova.exception.PortBindingFailed: Binding failed for port 52a0fa94-676e-4957-a12e-c9c1cd324b75, please check neutron logs for more information. [ 799.154757] env[62952]: ERROR nova.compute.manager [instance: da96503a-8ae8-4b68-a782-631ee98754f7] [ 799.154757] env[62952]: DEBUG nova.compute.utils [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Binding failed for port 52a0fa94-676e-4957-a12e-c9c1cd324b75, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 799.155471] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.131s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.159973] env[62952]: DEBUG nova.compute.manager [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Build of instance da96503a-8ae8-4b68-a782-631ee98754f7 was re-scheduled: Binding failed for port 52a0fa94-676e-4957-a12e-c9c1cd324b75, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 799.159973] env[62952]: DEBUG nova.compute.manager [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 799.159973] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Acquiring lock "refresh_cache-da96503a-8ae8-4b68-a782-631ee98754f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.159973] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Acquired lock "refresh_cache-da96503a-8ae8-4b68-a782-631ee98754f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.160256] env[62952]: DEBUG nova.network.neutron [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 799.198854] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.295697] env[62952]: DEBUG nova.compute.manager [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 799.318295] env[62952]: DEBUG nova.network.neutron [req-7da18bda-da8c-44fc-ae18-e43293a9ef7d req-4749ba80-e9d7-4a3f-aac4-d4a85f22d5ea service nova] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 799.321762] env[62952]: DEBUG nova.virt.hardware [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 799.322965] env[62952]: DEBUG nova.virt.hardware [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 799.322965] env[62952]: DEBUG nova.virt.hardware [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 799.322965] env[62952]: DEBUG nova.virt.hardware [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 799.322965] env[62952]: DEBUG nova.virt.hardware [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 799.322965] env[62952]: DEBUG nova.virt.hardware [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 799.323191] env[62952]: DEBUG nova.virt.hardware [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 799.323271] env[62952]: DEBUG nova.virt.hardware [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 799.323436] env[62952]: DEBUG nova.virt.hardware [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 799.323595] env[62952]: DEBUG nova.virt.hardware [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 799.323823] env[62952]: DEBUG nova.virt.hardware [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 799.325143] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc55a83-abcf-4ee9-a0de-b5f3a45f869f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.333809] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d9f8a90-c71a-4b69-8a98-ceb8542a7a0d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.347619] env[62952]: ERROR nova.compute.manager [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6629a780-8685-4991-9015-0601155c4978, please check neutron logs for more information. [ 799.347619] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Traceback (most recent call last): [ 799.347619] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 799.347619] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] yield resources [ 799.347619] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 799.347619] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] self.driver.spawn(context, instance, image_meta, [ 799.347619] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 799.347619] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 799.347619] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 799.347619] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] vm_ref = self.build_virtual_machine(instance, [ 799.347619] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 799.348081] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] vif_infos = vmwarevif.get_vif_info(self._session, [ 799.348081] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 799.348081] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] for vif in network_info: [ 799.348081] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 799.348081] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] return self._sync_wrapper(fn, *args, **kwargs) [ 799.348081] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 799.348081] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] self.wait() [ 799.348081] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 799.348081] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] self[:] = self._gt.wait() [ 799.348081] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 799.348081] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] return self._exit_event.wait() [ 799.348081] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 799.348081] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] current.throw(*self._exc) [ 799.348456] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 799.348456] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] result = function(*args, **kwargs) [ 799.348456] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 799.348456] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] return func(*args, **kwargs) [ 799.348456] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 799.348456] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] raise e [ 799.348456] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 799.348456] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] nwinfo = self.network_api.allocate_for_instance( [ 799.348456] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 799.348456] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] created_port_ids = self._update_ports_for_instance( [ 799.348456] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 799.348456] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] with excutils.save_and_reraise_exception(): [ 799.348456] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 799.348833] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] self.force_reraise() [ 799.348833] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 799.348833] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] raise self.value [ 799.348833] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 799.348833] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] updated_port = self._update_port( [ 799.348833] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 799.348833] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] _ensure_no_port_binding_failure(port) [ 799.348833] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 799.348833] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] raise exception.PortBindingFailed(port_id=port['id']) [ 799.348833] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] nova.exception.PortBindingFailed: Binding failed for port 6629a780-8685-4991-9015-0601155c4978, please check neutron logs for more information. [ 799.348833] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] [ 799.348833] env[62952]: INFO nova.compute.manager [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Terminating instance [ 799.353228] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Acquiring lock "refresh_cache-ebff0c87-8e58-4a76-a350-43a3d660debd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.394865] env[62952]: DEBUG nova.network.neutron [req-7da18bda-da8c-44fc-ae18-e43293a9ef7d req-4749ba80-e9d7-4a3f-aac4-d4a85f22d5ea service nova] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.679518] env[62952]: DEBUG nova.network.neutron [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 799.757193] env[62952]: DEBUG nova.network.neutron [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.897468] env[62952]: DEBUG oslo_concurrency.lockutils [req-7da18bda-da8c-44fc-ae18-e43293a9ef7d req-4749ba80-e9d7-4a3f-aac4-d4a85f22d5ea service nova] Releasing lock "refresh_cache-ebff0c87-8e58-4a76-a350-43a3d660debd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.897857] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Acquired lock "refresh_cache-ebff0c87-8e58-4a76-a350-43a3d660debd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.898046] env[62952]: DEBUG nova.network.neutron [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 799.950730] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ecee3e-9e28-45f2-92fa-41ffe06d0d8f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.958470] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44196b95-fde6-451c-ae2c-1057c081654d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.987180] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25ea4eb5-198d-43a9-9fdc-1479e30af7ae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.994064] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57938958-1780-4f98-bf43-45099aaf63ee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.007524] env[62952]: DEBUG nova.compute.provider_tree [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.260195] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Releasing lock "refresh_cache-da96503a-8ae8-4b68-a782-631ee98754f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.260469] env[62952]: DEBUG nova.compute.manager [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 800.260638] env[62952]: DEBUG nova.compute.manager [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 800.260803] env[62952]: DEBUG nova.network.neutron [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 800.275380] env[62952]: DEBUG nova.network.neutron [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 800.415158] env[62952]: DEBUG nova.network.neutron [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 800.484535] env[62952]: DEBUG nova.network.neutron [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.510883] env[62952]: DEBUG nova.scheduler.client.report [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 800.778627] env[62952]: DEBUG nova.network.neutron [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.877232] env[62952]: DEBUG nova.compute.manager [req-c9d9c314-5528-4b0e-8151-4a31f4be3a59 req-98d49716-9055-4483-9c0f-45673e8223ea service nova] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Received event network-vif-deleted-6629a780-8685-4991-9015-0601155c4978 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 800.987177] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Releasing lock "refresh_cache-ebff0c87-8e58-4a76-a350-43a3d660debd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.987603] env[62952]: DEBUG nova.compute.manager [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 800.987793] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 800.988110] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f2eac30c-0e07-4018-b174-a96d8f0489aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.997148] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03449659-cfa1-42f4-ba99-13194ae07984 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.018038] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.862s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.018672] env[62952]: ERROR nova.compute.manager [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 211aa71f-6364-48db-aeb2-5e51f549a03f, please check neutron logs for more information. [ 801.018672] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Traceback (most recent call last): [ 801.018672] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 801.018672] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] self.driver.spawn(context, instance, image_meta, [ 801.018672] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 801.018672] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] self._vmops.spawn(context, instance, image_meta, injected_files, [ 801.018672] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 801.018672] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] vm_ref = self.build_virtual_machine(instance, [ 801.018672] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 801.018672] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] vif_infos = vmwarevif.get_vif_info(self._session, [ 801.018672] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 801.019116] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] for vif in network_info: [ 801.019116] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 801.019116] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] return self._sync_wrapper(fn, *args, **kwargs) [ 801.019116] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 801.019116] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] self.wait() [ 801.019116] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 801.019116] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] self[:] = self._gt.wait() [ 801.019116] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 801.019116] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] return self._exit_event.wait() [ 801.019116] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 801.019116] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] current.throw(*self._exc) [ 801.019116] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 801.019116] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] result = function(*args, **kwargs) [ 801.019524] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 801.019524] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] return func(*args, **kwargs) [ 801.019524] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 801.019524] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] raise e [ 801.019524] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 801.019524] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] nwinfo = self.network_api.allocate_for_instance( [ 801.019524] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 801.019524] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] created_port_ids = self._update_ports_for_instance( [ 801.019524] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 801.019524] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] with excutils.save_and_reraise_exception(): [ 801.019524] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 801.019524] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] self.force_reraise() [ 801.019524] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 801.019949] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] raise self.value [ 801.019949] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 801.019949] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] updated_port = self._update_port( [ 801.019949] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 801.019949] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] _ensure_no_port_binding_failure(port) [ 801.019949] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 801.019949] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] raise exception.PortBindingFailed(port_id=port['id']) [ 801.019949] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] nova.exception.PortBindingFailed: Binding failed for port 211aa71f-6364-48db-aeb2-5e51f549a03f, please check neutron logs for more information. [ 801.019949] env[62952]: ERROR nova.compute.manager [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] [ 801.019949] env[62952]: DEBUG nova.compute.utils [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Binding failed for port 211aa71f-6364-48db-aeb2-5e51f549a03f, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 801.020673] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ebff0c87-8e58-4a76-a350-43a3d660debd could not be found. [ 801.020857] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 801.021061] env[62952]: INFO nova.compute.manager [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Took 0.03 seconds to destroy the instance on the hypervisor. [ 801.021265] env[62952]: DEBUG oslo.service.loopingcall [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 801.021710] env[62952]: DEBUG nova.compute.manager [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Build of instance 7f77fa00-27be-4a75-aac0-e1918f817837 was re-scheduled: Binding failed for port 211aa71f-6364-48db-aeb2-5e51f549a03f, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 801.022108] env[62952]: DEBUG nova.compute.manager [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 801.022325] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Acquiring lock "refresh_cache-7f77fa00-27be-4a75-aac0-e1918f817837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.022467] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Acquired lock "refresh_cache-7f77fa00-27be-4a75-aac0-e1918f817837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.022618] env[62952]: DEBUG nova.network.neutron [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 801.023548] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.619s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.024952] env[62952]: INFO nova.compute.claims [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 801.027211] env[62952]: DEBUG nova.compute.manager [-] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 801.027322] env[62952]: DEBUG nova.network.neutron [-] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 801.044523] env[62952]: DEBUG nova.network.neutron [-] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 801.282058] env[62952]: INFO nova.compute.manager [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] [instance: da96503a-8ae8-4b68-a782-631ee98754f7] Took 1.02 seconds to deallocate network for instance. [ 801.546566] env[62952]: DEBUG nova.network.neutron [-] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.548328] env[62952]: DEBUG nova.network.neutron [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 801.620284] env[62952]: DEBUG nova.network.neutron [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.051702] env[62952]: INFO nova.compute.manager [-] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Took 1.02 seconds to deallocate network for instance. [ 802.056590] env[62952]: DEBUG nova.compute.claims [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 802.056898] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.122532] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Releasing lock "refresh_cache-7f77fa00-27be-4a75-aac0-e1918f817837" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.122677] env[62952]: DEBUG nova.compute.manager [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 802.122859] env[62952]: DEBUG nova.compute.manager [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 802.123032] env[62952]: DEBUG nova.network.neutron [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 802.138221] env[62952]: DEBUG nova.network.neutron [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 802.307361] env[62952]: INFO nova.scheduler.client.report [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Deleted allocations for instance da96503a-8ae8-4b68-a782-631ee98754f7 [ 802.313770] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef09777e-12dd-4460-8d5e-29837806830b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.322926] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2945843a-d2f1-45de-b41f-c1b6ea16d630 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.354287] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c202f68-278f-44ec-8e9d-664ad9478e9c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.362355] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-444a062e-4d2d-4b60-a640-bd7470d30da7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.376464] env[62952]: DEBUG nova.compute.provider_tree [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.641127] env[62952]: DEBUG nova.network.neutron [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.819443] env[62952]: DEBUG oslo_concurrency.lockutils [None req-99a9efeb-ea67-469c-b9d5-6b832f629a0b tempest-ImagesOneServerTestJSON-488669050 tempest-ImagesOneServerTestJSON-488669050-project-member] Lock "da96503a-8ae8-4b68-a782-631ee98754f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.284s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.879029] env[62952]: DEBUG nova.scheduler.client.report [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 803.143696] env[62952]: INFO nova.compute.manager [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] [instance: 7f77fa00-27be-4a75-aac0-e1918f817837] Took 1.02 seconds to deallocate network for instance. [ 803.322123] env[62952]: DEBUG nova.compute.manager [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 803.384559] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.361s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.385076] env[62952]: DEBUG nova.compute.manager [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 803.388458] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.201s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.388458] env[62952]: DEBUG nova.objects.instance [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62952) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 803.843043] env[62952]: DEBUG oslo_concurrency.lockutils [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.897075] env[62952]: DEBUG nova.compute.utils [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.898594] env[62952]: DEBUG nova.compute.manager [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 803.898944] env[62952]: DEBUG nova.network.neutron [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 803.940924] env[62952]: DEBUG nova.policy [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'af4cc88249204af8bbf1fe63cd5e5a65', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ee886e2afe944cb29af54900efb88985', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 804.176384] env[62952]: INFO nova.scheduler.client.report [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Deleted allocations for instance 7f77fa00-27be-4a75-aac0-e1918f817837 [ 804.249788] env[62952]: DEBUG nova.network.neutron [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Successfully created port: f55e5aac-9df2-468a-8f4c-13216c743efc {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 804.400311] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3dcb8ad7-6916-4de9-84b3-ed3d7c490613 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.401459] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.208s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.406448] env[62952]: DEBUG nova.compute.manager [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 804.686384] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f491eb63-8296-499d-ab4c-edb11d0a4b89 tempest-InstanceActionsTestJSON-1418329997 tempest-InstanceActionsTestJSON-1418329997-project-member] Lock "7f77fa00-27be-4a75-aac0-e1918f817837" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.952s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.188167] env[62952]: DEBUG nova.compute.manager [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 805.256696] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-676c2940-cfdc-465a-a260-e5a8d57070b9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.267790] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eeced13-3a07-4c50-81b2-17b336573b81 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.301480] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc619caa-7856-4363-bb9c-641f10d9121e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.311951] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97324e87-9e92-41bc-aaa5-585d35d8e629 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.327094] env[62952]: DEBUG nova.compute.provider_tree [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.419373] env[62952]: DEBUG nova.compute.manager [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 805.435120] env[62952]: DEBUG nova.compute.manager [req-d6ec87b3-f7c2-448f-aac4-b7c905e346e1 req-d04ab1a7-a9b8-4c52-8ee9-1e60e95c16fc service nova] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Received event network-changed-f55e5aac-9df2-468a-8f4c-13216c743efc {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 805.435326] env[62952]: DEBUG nova.compute.manager [req-d6ec87b3-f7c2-448f-aac4-b7c905e346e1 req-d04ab1a7-a9b8-4c52-8ee9-1e60e95c16fc service nova] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Refreshing instance network info cache due to event network-changed-f55e5aac-9df2-468a-8f4c-13216c743efc. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 805.435539] env[62952]: DEBUG oslo_concurrency.lockutils [req-d6ec87b3-f7c2-448f-aac4-b7c905e346e1 req-d04ab1a7-a9b8-4c52-8ee9-1e60e95c16fc service nova] Acquiring lock "refresh_cache-1ed9de84-6178-4173-84ad-dc115be7a1fa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.435678] env[62952]: DEBUG oslo_concurrency.lockutils [req-d6ec87b3-f7c2-448f-aac4-b7c905e346e1 req-d04ab1a7-a9b8-4c52-8ee9-1e60e95c16fc service nova] Acquired lock "refresh_cache-1ed9de84-6178-4173-84ad-dc115be7a1fa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.435837] env[62952]: DEBUG nova.network.neutron [req-d6ec87b3-f7c2-448f-aac4-b7c905e346e1 req-d04ab1a7-a9b8-4c52-8ee9-1e60e95c16fc service nova] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Refreshing network info cache for port f55e5aac-9df2-468a-8f4c-13216c743efc {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 805.453250] env[62952]: DEBUG nova.virt.hardware [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 805.453626] env[62952]: DEBUG nova.virt.hardware [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 805.453715] env[62952]: DEBUG nova.virt.hardware [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 805.453836] env[62952]: DEBUG nova.virt.hardware [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 805.453977] env[62952]: DEBUG nova.virt.hardware [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 805.454371] env[62952]: DEBUG nova.virt.hardware [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 805.454620] env[62952]: DEBUG nova.virt.hardware [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 805.454782] env[62952]: DEBUG nova.virt.hardware [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 805.454948] env[62952]: DEBUG nova.virt.hardware [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 805.455125] env[62952]: DEBUG nova.virt.hardware [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 805.455303] env[62952]: DEBUG nova.virt.hardware [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 805.456574] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee495334-f9a1-4130-9574-7a8c1c1a88c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.465801] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-193ecd23-07a0-4f8f-bc32-7e03bcd64b8d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.609685] env[62952]: ERROR nova.compute.manager [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f55e5aac-9df2-468a-8f4c-13216c743efc, please check neutron logs for more information. [ 805.609685] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 805.609685] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 805.609685] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 805.609685] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 805.609685] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 805.609685] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 805.609685] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 805.609685] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 805.609685] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 805.609685] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 805.609685] env[62952]: ERROR nova.compute.manager raise self.value [ 805.609685] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 805.609685] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 805.609685] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 805.609685] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 805.610230] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 805.610230] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 805.610230] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f55e5aac-9df2-468a-8f4c-13216c743efc, please check neutron logs for more information. [ 805.610230] env[62952]: ERROR nova.compute.manager [ 805.610230] env[62952]: Traceback (most recent call last): [ 805.610230] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 805.610230] env[62952]: listener.cb(fileno) [ 805.610230] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 805.610230] env[62952]: result = function(*args, **kwargs) [ 805.610230] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 805.610230] env[62952]: return func(*args, **kwargs) [ 805.610230] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 805.610230] env[62952]: raise e [ 805.610230] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 805.610230] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 805.610230] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 805.610230] env[62952]: created_port_ids = self._update_ports_for_instance( [ 805.610230] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 805.610230] env[62952]: with excutils.save_and_reraise_exception(): [ 805.610230] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 805.610230] env[62952]: self.force_reraise() [ 805.610230] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 805.610230] env[62952]: raise self.value [ 805.610230] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 805.610230] env[62952]: updated_port = self._update_port( [ 805.610230] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 805.610230] env[62952]: _ensure_no_port_binding_failure(port) [ 805.610230] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 805.610230] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 805.611116] env[62952]: nova.exception.PortBindingFailed: Binding failed for port f55e5aac-9df2-468a-8f4c-13216c743efc, please check neutron logs for more information. [ 805.611116] env[62952]: Removing descriptor: 19 [ 805.611116] env[62952]: ERROR nova.compute.manager [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f55e5aac-9df2-468a-8f4c-13216c743efc, please check neutron logs for more information. [ 805.611116] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Traceback (most recent call last): [ 805.611116] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 805.611116] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] yield resources [ 805.611116] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 805.611116] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] self.driver.spawn(context, instance, image_meta, [ 805.611116] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 805.611116] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 805.611116] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 805.611116] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] vm_ref = self.build_virtual_machine(instance, [ 805.611492] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 805.611492] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] vif_infos = vmwarevif.get_vif_info(self._session, [ 805.611492] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 805.611492] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] for vif in network_info: [ 805.611492] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 805.611492] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] return self._sync_wrapper(fn, *args, **kwargs) [ 805.611492] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 805.611492] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] self.wait() [ 805.611492] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 805.611492] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] self[:] = self._gt.wait() [ 805.611492] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 805.611492] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] return self._exit_event.wait() [ 805.611492] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 805.611888] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] result = hub.switch() [ 805.611888] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 805.611888] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] return self.greenlet.switch() [ 805.611888] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 805.611888] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] result = function(*args, **kwargs) [ 805.611888] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 805.611888] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] return func(*args, **kwargs) [ 805.611888] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 805.611888] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] raise e [ 805.611888] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 805.611888] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] nwinfo = self.network_api.allocate_for_instance( [ 805.611888] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 805.611888] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] created_port_ids = self._update_ports_for_instance( [ 805.612316] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 805.612316] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] with excutils.save_and_reraise_exception(): [ 805.612316] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 805.612316] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] self.force_reraise() [ 805.612316] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 805.612316] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] raise self.value [ 805.612316] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 805.612316] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] updated_port = self._update_port( [ 805.612316] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 805.612316] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] _ensure_no_port_binding_failure(port) [ 805.612316] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 805.612316] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] raise exception.PortBindingFailed(port_id=port['id']) [ 805.612693] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] nova.exception.PortBindingFailed: Binding failed for port f55e5aac-9df2-468a-8f4c-13216c743efc, please check neutron logs for more information. [ 805.612693] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] [ 805.612693] env[62952]: INFO nova.compute.manager [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Terminating instance [ 805.615995] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Acquiring lock "refresh_cache-1ed9de84-6178-4173-84ad-dc115be7a1fa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.722067] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.831696] env[62952]: DEBUG nova.scheduler.client.report [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 805.957664] env[62952]: DEBUG nova.network.neutron [req-d6ec87b3-f7c2-448f-aac4-b7c905e346e1 req-d04ab1a7-a9b8-4c52-8ee9-1e60e95c16fc service nova] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.046139] env[62952]: DEBUG nova.network.neutron [req-d6ec87b3-f7c2-448f-aac4-b7c905e346e1 req-d04ab1a7-a9b8-4c52-8ee9-1e60e95c16fc service nova] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.338313] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.937s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.338951] env[62952]: ERROR nova.compute.manager [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 84777b2a-9e38-4048-b45c-f263fb473613, please check neutron logs for more information. [ 806.338951] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Traceback (most recent call last): [ 806.338951] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 806.338951] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] self.driver.spawn(context, instance, image_meta, [ 806.338951] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 806.338951] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] self._vmops.spawn(context, instance, image_meta, injected_files, [ 806.338951] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 806.338951] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] vm_ref = self.build_virtual_machine(instance, [ 806.338951] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 806.338951] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] vif_infos = vmwarevif.get_vif_info(self._session, [ 806.338951] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 806.339314] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] for vif in network_info: [ 806.339314] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 806.339314] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] return self._sync_wrapper(fn, *args, **kwargs) [ 806.339314] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 806.339314] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] self.wait() [ 806.339314] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 806.339314] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] self[:] = self._gt.wait() [ 806.339314] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 806.339314] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] return self._exit_event.wait() [ 806.339314] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 806.339314] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] current.throw(*self._exc) [ 806.339314] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 806.339314] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] result = function(*args, **kwargs) [ 806.339762] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 806.339762] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] return func(*args, **kwargs) [ 806.339762] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 806.339762] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] raise e [ 806.339762] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 806.339762] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] nwinfo = self.network_api.allocate_for_instance( [ 806.339762] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 806.339762] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] created_port_ids = self._update_ports_for_instance( [ 806.339762] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 806.339762] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] with excutils.save_and_reraise_exception(): [ 806.339762] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 806.339762] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] self.force_reraise() [ 806.339762] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 806.340116] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] raise self.value [ 806.340116] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 806.340116] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] updated_port = self._update_port( [ 806.340116] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 806.340116] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] _ensure_no_port_binding_failure(port) [ 806.340116] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 806.340116] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] raise exception.PortBindingFailed(port_id=port['id']) [ 806.340116] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] nova.exception.PortBindingFailed: Binding failed for port 84777b2a-9e38-4048-b45c-f263fb473613, please check neutron logs for more information. [ 806.340116] env[62952]: ERROR nova.compute.manager [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] [ 806.340116] env[62952]: DEBUG nova.compute.utils [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Binding failed for port 84777b2a-9e38-4048-b45c-f263fb473613, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 806.340919] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.108s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.342803] env[62952]: INFO nova.compute.claims [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.345879] env[62952]: DEBUG nova.compute.manager [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Build of instance 879dc564-0a5e-48aa-92ff-b02435445a68 was re-scheduled: Binding failed for port 84777b2a-9e38-4048-b45c-f263fb473613, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 806.346317] env[62952]: DEBUG nova.compute.manager [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 806.346542] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Acquiring lock "refresh_cache-879dc564-0a5e-48aa-92ff-b02435445a68" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.346840] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Acquired lock "refresh_cache-879dc564-0a5e-48aa-92ff-b02435445a68" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.347027] env[62952]: DEBUG nova.network.neutron [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 806.549339] env[62952]: DEBUG oslo_concurrency.lockutils [req-d6ec87b3-f7c2-448f-aac4-b7c905e346e1 req-d04ab1a7-a9b8-4c52-8ee9-1e60e95c16fc service nova] Releasing lock "refresh_cache-1ed9de84-6178-4173-84ad-dc115be7a1fa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.549820] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Acquired lock "refresh_cache-1ed9de84-6178-4173-84ad-dc115be7a1fa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.550014] env[62952]: DEBUG nova.network.neutron [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 806.868392] env[62952]: DEBUG nova.network.neutron [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.995807] env[62952]: DEBUG nova.network.neutron [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.067937] env[62952]: DEBUG nova.network.neutron [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 807.145611] env[62952]: DEBUG nova.network.neutron [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.464215] env[62952]: DEBUG nova.compute.manager [req-bea7ee8d-694e-4052-b2f6-257dc77af0e8 req-46187c70-46ee-4a2f-b722-d5b98f97b564 service nova] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Received event network-vif-deleted-f55e5aac-9df2-468a-8f4c-13216c743efc {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 807.502634] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Releasing lock "refresh_cache-879dc564-0a5e-48aa-92ff-b02435445a68" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.502634] env[62952]: DEBUG nova.compute.manager [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 807.502634] env[62952]: DEBUG nova.compute.manager [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 807.502634] env[62952]: DEBUG nova.network.neutron [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 807.518339] env[62952]: DEBUG nova.network.neutron [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 807.635020] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b55c3081-ec61-43fc-8b0d-485c9fc31a87 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.642724] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5137e0-4f86-46ff-b795-62e57f9ec742 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.648453] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Releasing lock "refresh_cache-1ed9de84-6178-4173-84ad-dc115be7a1fa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.648850] env[62952]: DEBUG nova.compute.manager [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 807.649048] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 807.673895] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-29e011e0-388e-4ce5-acc0-edcf595edfe2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.676752] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9a69c8-5810-49c8-88fc-0c97023a66a0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.687409] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec85951-8d0c-48cb-9667-731ea496fec1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.693577] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bafa36bb-8733-4e8b-9b88-1dc3c006a23d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.713873] env[62952]: DEBUG nova.compute.provider_tree [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.722021] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1ed9de84-6178-4173-84ad-dc115be7a1fa could not be found. [ 807.722021] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 807.722021] env[62952]: INFO nova.compute.manager [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Took 0.07 seconds to destroy the instance on the hypervisor. [ 807.722021] env[62952]: DEBUG oslo.service.loopingcall [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.722021] env[62952]: DEBUG nova.compute.manager [-] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 807.722021] env[62952]: DEBUG nova.network.neutron [-] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 807.740332] env[62952]: DEBUG nova.network.neutron [-] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 808.023894] env[62952]: DEBUG nova.network.neutron [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.221094] env[62952]: DEBUG nova.scheduler.client.report [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.243849] env[62952]: DEBUG nova.network.neutron [-] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.527741] env[62952]: INFO nova.compute.manager [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] [instance: 879dc564-0a5e-48aa-92ff-b02435445a68] Took 1.03 seconds to deallocate network for instance. [ 808.725898] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.726468] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 808.729033] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.923s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.746176] env[62952]: INFO nova.compute.manager [-] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Took 1.03 seconds to deallocate network for instance. [ 808.748336] env[62952]: DEBUG nova.compute.claims [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 808.748521] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.235029] env[62952]: DEBUG nova.compute.utils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 809.242253] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 809.242253] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 809.297294] env[62952]: DEBUG nova.policy [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8fdf784c61c44acc87d6fede9b4b150f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eac1e69a14e94fb89ca59c8d0b5edfaf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 809.569783] env[62952]: INFO nova.scheduler.client.report [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Deleted allocations for instance 879dc564-0a5e-48aa-92ff-b02435445a68 [ 809.579591] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00d24bcb-7877-44c4-8e6b-12c1f2e64869 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.589406] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a25e5e1-a84e-492a-a898-e82f01af4487 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.595426] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Successfully created port: 559b10ed-15b2-417c-bf21-939af6239155 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 809.628487] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce07377-1a92-456a-be52-866a11b6ce2a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.636804] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dccf8ce8-4d61-4839-b8b5-a2c2ca32efd1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.651341] env[62952]: DEBUG nova.compute.provider_tree [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.744100] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 810.085221] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8dd11687-ba34-4ad4-b20a-6f6a52b1461a tempest-SecurityGroupsTestJSON-689853685 tempest-SecurityGroupsTestJSON-689853685-project-member] Lock "879dc564-0a5e-48aa-92ff-b02435445a68" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.428s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.154729] env[62952]: DEBUG nova.scheduler.client.report [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 810.482160] env[62952]: DEBUG nova.compute.manager [req-8c30fde8-9799-48ae-9056-3a377985ed84 req-510e7399-ac19-478e-8121-bb6f0136c19c service nova] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Received event network-changed-559b10ed-15b2-417c-bf21-939af6239155 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 810.482561] env[62952]: DEBUG nova.compute.manager [req-8c30fde8-9799-48ae-9056-3a377985ed84 req-510e7399-ac19-478e-8121-bb6f0136c19c service nova] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Refreshing instance network info cache due to event network-changed-559b10ed-15b2-417c-bf21-939af6239155. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 810.482961] env[62952]: DEBUG oslo_concurrency.lockutils [req-8c30fde8-9799-48ae-9056-3a377985ed84 req-510e7399-ac19-478e-8121-bb6f0136c19c service nova] Acquiring lock "refresh_cache-1d6447ab-dc07-459f-9124-d881e0da7cf1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.484790] env[62952]: DEBUG oslo_concurrency.lockutils [req-8c30fde8-9799-48ae-9056-3a377985ed84 req-510e7399-ac19-478e-8121-bb6f0136c19c service nova] Acquired lock "refresh_cache-1d6447ab-dc07-459f-9124-d881e0da7cf1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.484790] env[62952]: DEBUG nova.network.neutron [req-8c30fde8-9799-48ae-9056-3a377985ed84 req-510e7399-ac19-478e-8121-bb6f0136c19c service nova] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Refreshing network info cache for port 559b10ed-15b2-417c-bf21-939af6239155 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 810.587579] env[62952]: DEBUG nova.compute.manager [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 810.659667] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.930s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.660370] env[62952]: ERROR nova.compute.manager [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port aa429812-d328-4234-92d8-01fbd7263815, please check neutron logs for more information. [ 810.660370] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Traceback (most recent call last): [ 810.660370] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 810.660370] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] self.driver.spawn(context, instance, image_meta, [ 810.660370] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 810.660370] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 810.660370] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 810.660370] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] vm_ref = self.build_virtual_machine(instance, [ 810.660370] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 810.660370] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] vif_infos = vmwarevif.get_vif_info(self._session, [ 810.660370] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 810.660722] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] for vif in network_info: [ 810.660722] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 810.660722] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] return self._sync_wrapper(fn, *args, **kwargs) [ 810.660722] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 810.660722] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] self.wait() [ 810.660722] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 810.660722] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] self[:] = self._gt.wait() [ 810.660722] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 810.660722] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] return self._exit_event.wait() [ 810.660722] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 810.660722] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] current.throw(*self._exc) [ 810.660722] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 810.660722] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] result = function(*args, **kwargs) [ 810.661082] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 810.661082] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] return func(*args, **kwargs) [ 810.661082] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 810.661082] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] raise e [ 810.661082] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 810.661082] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] nwinfo = self.network_api.allocate_for_instance( [ 810.661082] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 810.661082] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] created_port_ids = self._update_ports_for_instance( [ 810.661082] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 810.661082] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] with excutils.save_and_reraise_exception(): [ 810.661082] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 810.661082] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] self.force_reraise() [ 810.661082] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 810.661430] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] raise self.value [ 810.661430] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 810.661430] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] updated_port = self._update_port( [ 810.661430] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 810.661430] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] _ensure_no_port_binding_failure(port) [ 810.661430] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 810.661430] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] raise exception.PortBindingFailed(port_id=port['id']) [ 810.661430] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] nova.exception.PortBindingFailed: Binding failed for port aa429812-d328-4234-92d8-01fbd7263815, please check neutron logs for more information. [ 810.661430] env[62952]: ERROR nova.compute.manager [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] [ 810.661430] env[62952]: DEBUG nova.compute.utils [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Binding failed for port aa429812-d328-4234-92d8-01fbd7263815, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 810.662337] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.747s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.666035] env[62952]: DEBUG nova.compute.manager [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Build of instance 3822c92d-560b-4d61-92e9-7c694bc7a33a was re-scheduled: Binding failed for port aa429812-d328-4234-92d8-01fbd7263815, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 810.666553] env[62952]: DEBUG nova.compute.manager [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 810.666886] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Acquiring lock "refresh_cache-3822c92d-560b-4d61-92e9-7c694bc7a33a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.667073] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Acquired lock "refresh_cache-3822c92d-560b-4d61-92e9-7c694bc7a33a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.667241] env[62952]: DEBUG nova.network.neutron [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 810.684891] env[62952]: ERROR nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 559b10ed-15b2-417c-bf21-939af6239155, please check neutron logs for more information. [ 810.684891] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 810.684891] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 810.684891] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 810.684891] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 810.684891] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 810.684891] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 810.684891] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 810.684891] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 810.684891] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 810.684891] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 810.684891] env[62952]: ERROR nova.compute.manager raise self.value [ 810.684891] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 810.684891] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 810.684891] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 810.684891] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 810.685666] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 810.685666] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 810.685666] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 559b10ed-15b2-417c-bf21-939af6239155, please check neutron logs for more information. [ 810.685666] env[62952]: ERROR nova.compute.manager [ 810.685666] env[62952]: Traceback (most recent call last): [ 810.685666] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 810.685666] env[62952]: listener.cb(fileno) [ 810.685666] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 810.685666] env[62952]: result = function(*args, **kwargs) [ 810.685666] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 810.685666] env[62952]: return func(*args, **kwargs) [ 810.685666] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 810.685666] env[62952]: raise e [ 810.685666] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 810.685666] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 810.685666] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 810.685666] env[62952]: created_port_ids = self._update_ports_for_instance( [ 810.685666] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 810.685666] env[62952]: with excutils.save_and_reraise_exception(): [ 810.685666] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 810.685666] env[62952]: self.force_reraise() [ 810.685666] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 810.685666] env[62952]: raise self.value [ 810.685666] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 810.685666] env[62952]: updated_port = self._update_port( [ 810.685666] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 810.685666] env[62952]: _ensure_no_port_binding_failure(port) [ 810.685666] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 810.685666] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 810.686950] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 559b10ed-15b2-417c-bf21-939af6239155, please check neutron logs for more information. [ 810.686950] env[62952]: Removing descriptor: 19 [ 810.755191] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 810.780956] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 810.781237] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 810.781500] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 810.781635] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 810.781784] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 810.781932] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 810.782183] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 810.782360] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 810.782594] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 810.782702] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 810.782862] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 810.783761] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d28fc440-4a49-4d18-895f-2a05e01069d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.791911] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd9cc20-fba6-4d7f-a134-976cd9ba7a8f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.805332] env[62952]: ERROR nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 559b10ed-15b2-417c-bf21-939af6239155, please check neutron logs for more information. [ 810.805332] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Traceback (most recent call last): [ 810.805332] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 810.805332] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] yield resources [ 810.805332] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 810.805332] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] self.driver.spawn(context, instance, image_meta, [ 810.805332] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 810.805332] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 810.805332] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 810.805332] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] vm_ref = self.build_virtual_machine(instance, [ 810.805332] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 810.805782] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] vif_infos = vmwarevif.get_vif_info(self._session, [ 810.805782] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 810.805782] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] for vif in network_info: [ 810.805782] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 810.805782] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] return self._sync_wrapper(fn, *args, **kwargs) [ 810.805782] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 810.805782] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] self.wait() [ 810.805782] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 810.805782] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] self[:] = self._gt.wait() [ 810.805782] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 810.805782] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] return self._exit_event.wait() [ 810.805782] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 810.805782] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] current.throw(*self._exc) [ 810.806166] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 810.806166] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] result = function(*args, **kwargs) [ 810.806166] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 810.806166] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] return func(*args, **kwargs) [ 810.806166] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 810.806166] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] raise e [ 810.806166] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 810.806166] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] nwinfo = self.network_api.allocate_for_instance( [ 810.806166] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 810.806166] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] created_port_ids = self._update_ports_for_instance( [ 810.806166] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 810.806166] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] with excutils.save_and_reraise_exception(): [ 810.806166] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 810.806530] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] self.force_reraise() [ 810.806530] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 810.806530] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] raise self.value [ 810.806530] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 810.806530] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] updated_port = self._update_port( [ 810.806530] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 810.806530] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] _ensure_no_port_binding_failure(port) [ 810.806530] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 810.806530] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] raise exception.PortBindingFailed(port_id=port['id']) [ 810.806530] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] nova.exception.PortBindingFailed: Binding failed for port 559b10ed-15b2-417c-bf21-939af6239155, please check neutron logs for more information. [ 810.806530] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] [ 810.806530] env[62952]: INFO nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Terminating instance [ 810.807723] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "refresh_cache-1d6447ab-dc07-459f-9124-d881e0da7cf1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.001851] env[62952]: DEBUG nova.network.neutron [req-8c30fde8-9799-48ae-9056-3a377985ed84 req-510e7399-ac19-478e-8121-bb6f0136c19c service nova] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 811.091701] env[62952]: DEBUG nova.network.neutron [req-8c30fde8-9799-48ae-9056-3a377985ed84 req-510e7399-ac19-478e-8121-bb6f0136c19c service nova] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.119177] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.189546] env[62952]: DEBUG nova.network.neutron [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 811.278088] env[62952]: DEBUG nova.network.neutron [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.441738] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc1d274b-4f5d-44f5-8773-8c403f7b4e68 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.449466] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b09236d8-6eb4-4f5a-973b-0a3cda4548e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.481017] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54470a47-5430-400c-9e98-d4b6e0fa8492 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.488099] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bacf0fb-0f52-46e9-8787-e5dd6d40d13e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.501214] env[62952]: DEBUG nova.compute.provider_tree [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.593883] env[62952]: DEBUG oslo_concurrency.lockutils [req-8c30fde8-9799-48ae-9056-3a377985ed84 req-510e7399-ac19-478e-8121-bb6f0136c19c service nova] Releasing lock "refresh_cache-1d6447ab-dc07-459f-9124-d881e0da7cf1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.594331] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquired lock "refresh_cache-1d6447ab-dc07-459f-9124-d881e0da7cf1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.595219] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 811.783728] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Releasing lock "refresh_cache-3822c92d-560b-4d61-92e9-7c694bc7a33a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.783728] env[62952]: DEBUG nova.compute.manager [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 811.783728] env[62952]: DEBUG nova.compute.manager [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 811.783728] env[62952]: DEBUG nova.network.neutron [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 811.798439] env[62952]: DEBUG nova.network.neutron [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 812.007017] env[62952]: DEBUG nova.scheduler.client.report [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 812.116254] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 812.240689] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.299636] env[62952]: DEBUG nova.network.neutron [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.507180] env[62952]: DEBUG nova.compute.manager [req-9767d961-59d5-441e-8ac2-db5bfadfeaa4 req-c2320431-23f3-4e0c-8f15-de4c23e60999 service nova] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Received event network-vif-deleted-559b10ed-15b2-417c-bf21-939af6239155 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 812.510168] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.848s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.510816] env[62952]: ERROR nova.compute.manager [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 57491018-5fb5-4504-a04f-c8729d155e31, please check neutron logs for more information. [ 812.510816] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Traceback (most recent call last): [ 812.510816] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 812.510816] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] self.driver.spawn(context, instance, image_meta, [ 812.510816] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 812.510816] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] self._vmops.spawn(context, instance, image_meta, injected_files, [ 812.510816] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 812.510816] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] vm_ref = self.build_virtual_machine(instance, [ 812.510816] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 812.510816] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] vif_infos = vmwarevif.get_vif_info(self._session, [ 812.510816] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 812.511352] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] for vif in network_info: [ 812.511352] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 812.511352] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] return self._sync_wrapper(fn, *args, **kwargs) [ 812.511352] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 812.511352] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] self.wait() [ 812.511352] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 812.511352] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] self[:] = self._gt.wait() [ 812.511352] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 812.511352] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] return self._exit_event.wait() [ 812.511352] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 812.511352] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] current.throw(*self._exc) [ 812.511352] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 812.511352] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] result = function(*args, **kwargs) [ 812.511889] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 812.511889] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] return func(*args, **kwargs) [ 812.511889] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 812.511889] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] raise e [ 812.511889] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 812.511889] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] nwinfo = self.network_api.allocate_for_instance( [ 812.511889] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 812.511889] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] created_port_ids = self._update_ports_for_instance( [ 812.511889] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 812.511889] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] with excutils.save_and_reraise_exception(): [ 812.511889] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 812.511889] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] self.force_reraise() [ 812.511889] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 812.512259] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] raise self.value [ 812.512259] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 812.512259] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] updated_port = self._update_port( [ 812.512259] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 812.512259] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] _ensure_no_port_binding_failure(port) [ 812.512259] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 812.512259] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] raise exception.PortBindingFailed(port_id=port['id']) [ 812.512259] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] nova.exception.PortBindingFailed: Binding failed for port 57491018-5fb5-4504-a04f-c8729d155e31, please check neutron logs for more information. [ 812.512259] env[62952]: ERROR nova.compute.manager [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] [ 812.512259] env[62952]: DEBUG nova.compute.utils [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Binding failed for port 57491018-5fb5-4504-a04f-c8729d155e31, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 812.512976] env[62952]: DEBUG nova.compute.manager [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Build of instance 9979dd1d-9e99-47cf-a81a-fd7d89531f51 was re-scheduled: Binding failed for port 57491018-5fb5-4504-a04f-c8729d155e31, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 812.513664] env[62952]: DEBUG nova.compute.manager [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 812.513951] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Acquiring lock "refresh_cache-9979dd1d-9e99-47cf-a81a-fd7d89531f51" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.514156] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Acquired lock "refresh_cache-9979dd1d-9e99-47cf-a81a-fd7d89531f51" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.514365] env[62952]: DEBUG nova.network.neutron [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 812.515315] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.620s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.519255] env[62952]: DEBUG nova.objects.instance [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62952) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 812.747072] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Releasing lock "refresh_cache-1d6447ab-dc07-459f-9124-d881e0da7cf1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.747512] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 812.747766] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 812.748073] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e96433ed-d685-4f45-9b4c-de3fd6acbdf8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.757565] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4443ad2d-f5c9-4d20-b9eb-8efe0d30b4f7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.779299] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1d6447ab-dc07-459f-9124-d881e0da7cf1 could not be found. [ 812.779526] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 812.779711] env[62952]: INFO nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 812.779944] env[62952]: DEBUG oslo.service.loopingcall [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 812.780164] env[62952]: DEBUG nova.compute.manager [-] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 812.780259] env[62952]: DEBUG nova.network.neutron [-] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 812.795644] env[62952]: DEBUG nova.network.neutron [-] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 812.802775] env[62952]: INFO nova.compute.manager [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] [instance: 3822c92d-560b-4d61-92e9-7c694bc7a33a] Took 1.02 seconds to deallocate network for instance. [ 813.041306] env[62952]: DEBUG nova.network.neutron [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 813.095962] env[62952]: DEBUG nova.network.neutron [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.298173] env[62952]: DEBUG nova.network.neutron [-] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.530281] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0217c229-7057-4aab-bb9d-ba0c6be81125 tempest-ServersAdmin275Test-435701955 tempest-ServersAdmin275Test-435701955-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.531470] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.906s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.532945] env[62952]: INFO nova.compute.claims [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 813.598744] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Releasing lock "refresh_cache-9979dd1d-9e99-47cf-a81a-fd7d89531f51" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.599011] env[62952]: DEBUG nova.compute.manager [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 813.599215] env[62952]: DEBUG nova.compute.manager [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 813.599336] env[62952]: DEBUG nova.network.neutron [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 813.615678] env[62952]: DEBUG nova.network.neutron [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 813.801587] env[62952]: INFO nova.compute.manager [-] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Took 1.02 seconds to deallocate network for instance. [ 813.803974] env[62952]: DEBUG nova.compute.claims [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 813.804195] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.838646] env[62952]: INFO nova.scheduler.client.report [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Deleted allocations for instance 3822c92d-560b-4d61-92e9-7c694bc7a33a [ 814.118271] env[62952]: DEBUG nova.network.neutron [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.348996] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77529bd1-91fe-4c68-9933-50f6e7adb25f tempest-ServersTestBootFromVolume-1047562459 tempest-ServersTestBootFromVolume-1047562459-project-member] Lock "3822c92d-560b-4d61-92e9-7c694bc7a33a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.175s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.620902] env[62952]: INFO nova.compute.manager [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] [instance: 9979dd1d-9e99-47cf-a81a-fd7d89531f51] Took 1.02 seconds to deallocate network for instance. [ 814.781929] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 814.785339] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 814.845477] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7b11d99-e90c-47e7-b311-6ca8e9d60bfa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.848687] env[62952]: DEBUG nova.compute.manager [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 814.858352] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef7b6a3d-d6a1-4557-bd72-1e5b875443d9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.893246] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44ae2dd7-e3ec-41d2-8472-a24cc2d58c4a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.901329] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20bb326-312c-40af-88d7-3a8e0e82c833 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.916163] env[62952]: DEBUG nova.compute.provider_tree [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.292959] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 815.296251] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 815.296251] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 815.374494] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.419139] env[62952]: DEBUG nova.scheduler.client.report [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.662935] env[62952]: INFO nova.scheduler.client.report [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Deleted allocations for instance 9979dd1d-9e99-47cf-a81a-fd7d89531f51 [ 815.803021] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 815.803021] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 815.803021] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 815.803021] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 815.822826] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "refresh_cache-472e60f1-9b6c-415e-8a92-19235ad2e547" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.822980] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquired lock "refresh_cache-472e60f1-9b6c-415e-8a92-19235ad2e547" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.823142] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Forcefully refreshing network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 815.823301] env[62952]: DEBUG nova.objects.instance [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lazy-loading 'info_cache' on Instance uuid 472e60f1-9b6c-415e-8a92-19235ad2e547 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 815.924776] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.925330] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 815.928555] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.458s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.935028] env[62952]: INFO nova.compute.claims [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.173849] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6148896b-ccf8-4b2f-a0d8-e8c6c014f9ac tempest-MigrationsAdminTest-2079194532 tempest-MigrationsAdminTest-2079194532-project-member] Lock "9979dd1d-9e99-47cf-a81a-fd7d89531f51" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.013s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.440097] env[62952]: DEBUG nova.compute.utils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 816.443559] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 816.443718] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 816.498825] env[62952]: DEBUG nova.policy [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8fdf784c61c44acc87d6fede9b4b150f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eac1e69a14e94fb89ca59c8d0b5edfaf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 816.679775] env[62952]: DEBUG nova.compute.manager [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 816.787412] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "3416eb38-b961-4d48-8b37-8b7b92375dc9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.787741] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "3416eb38-b961-4d48-8b37-8b7b92375dc9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.849561] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 816.857382] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Successfully created port: c5ae85f8-f1b8-497a-a7b9-a9b53a3c813c {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 816.944497] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 817.210298] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.284386] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f94372c-70c7-4a94-8758-29e25980fc8c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.296413] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f3d49cc-35d6-4ef0-a730-b95439f3343d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.332847] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2674df3a-fbbd-4119-b41e-0e85ff16c6cd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.341469] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ef631a-4be3-4c52-8a28-2210845dbc5b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.361022] env[62952]: DEBUG nova.compute.provider_tree [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.444841] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.706155] env[62952]: DEBUG nova.compute.manager [req-dd138084-a895-4c7c-a2ba-919f06fff5c0 req-344aff31-8270-4110-8d26-2877e6ba3115 service nova] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Received event network-changed-c5ae85f8-f1b8-497a-a7b9-a9b53a3c813c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 817.706155] env[62952]: DEBUG nova.compute.manager [req-dd138084-a895-4c7c-a2ba-919f06fff5c0 req-344aff31-8270-4110-8d26-2877e6ba3115 service nova] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Refreshing instance network info cache due to event network-changed-c5ae85f8-f1b8-497a-a7b9-a9b53a3c813c. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 817.706155] env[62952]: DEBUG oslo_concurrency.lockutils [req-dd138084-a895-4c7c-a2ba-919f06fff5c0 req-344aff31-8270-4110-8d26-2877e6ba3115 service nova] Acquiring lock "refresh_cache-9c5167eb-3a15-47ff-923e-c83cc3529cc8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.706155] env[62952]: DEBUG oslo_concurrency.lockutils [req-dd138084-a895-4c7c-a2ba-919f06fff5c0 req-344aff31-8270-4110-8d26-2877e6ba3115 service nova] Acquired lock "refresh_cache-9c5167eb-3a15-47ff-923e-c83cc3529cc8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.706155] env[62952]: DEBUG nova.network.neutron [req-dd138084-a895-4c7c-a2ba-919f06fff5c0 req-344aff31-8270-4110-8d26-2877e6ba3115 service nova] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Refreshing network info cache for port c5ae85f8-f1b8-497a-a7b9-a9b53a3c813c {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 817.861372] env[62952]: DEBUG nova.scheduler.client.report [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 817.949763] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Releasing lock "refresh_cache-472e60f1-9b6c-415e-8a92-19235ad2e547" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.949763] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Updated the network info_cache for instance {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 817.949763] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 817.949763] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 817.949763] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 817.949763] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 817.950249] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 817.950249] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 817.950249] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 817.950249] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 817.957878] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 817.985173] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 817.985173] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 817.985173] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 817.985336] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 817.985336] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 817.985336] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 817.985336] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 817.985336] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 817.985506] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 817.985621] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 817.985928] env[62952]: DEBUG nova.virt.hardware [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 817.987011] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8031ac0-5a41-47bd-8807-59f68a20f1a0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.995442] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77d9ebcf-d317-4572-acfe-9acd81118787 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.038138] env[62952]: ERROR nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c5ae85f8-f1b8-497a-a7b9-a9b53a3c813c, please check neutron logs for more information. [ 818.038138] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 818.038138] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 818.038138] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 818.038138] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 818.038138] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 818.038138] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 818.038138] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 818.038138] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 818.038138] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 818.038138] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 818.038138] env[62952]: ERROR nova.compute.manager raise self.value [ 818.038138] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 818.038138] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 818.038138] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 818.038138] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 818.038676] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 818.038676] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 818.038676] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c5ae85f8-f1b8-497a-a7b9-a9b53a3c813c, please check neutron logs for more information. [ 818.038676] env[62952]: ERROR nova.compute.manager [ 818.038676] env[62952]: Traceback (most recent call last): [ 818.038676] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 818.038676] env[62952]: listener.cb(fileno) [ 818.038676] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 818.038676] env[62952]: result = function(*args, **kwargs) [ 818.038676] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 818.038676] env[62952]: return func(*args, **kwargs) [ 818.038676] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 818.038676] env[62952]: raise e [ 818.038676] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 818.038676] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 818.038676] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 818.038676] env[62952]: created_port_ids = self._update_ports_for_instance( [ 818.038676] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 818.038676] env[62952]: with excutils.save_and_reraise_exception(): [ 818.038676] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 818.038676] env[62952]: self.force_reraise() [ 818.038676] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 818.038676] env[62952]: raise self.value [ 818.038676] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 818.038676] env[62952]: updated_port = self._update_port( [ 818.038676] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 818.038676] env[62952]: _ensure_no_port_binding_failure(port) [ 818.038676] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 818.038676] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 818.039594] env[62952]: nova.exception.PortBindingFailed: Binding failed for port c5ae85f8-f1b8-497a-a7b9-a9b53a3c813c, please check neutron logs for more information. [ 818.039594] env[62952]: Removing descriptor: 19 [ 818.039594] env[62952]: ERROR nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c5ae85f8-f1b8-497a-a7b9-a9b53a3c813c, please check neutron logs for more information. [ 818.039594] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Traceback (most recent call last): [ 818.039594] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 818.039594] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] yield resources [ 818.039594] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 818.039594] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] self.driver.spawn(context, instance, image_meta, [ 818.039594] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 818.039594] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 818.039594] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 818.039594] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] vm_ref = self.build_virtual_machine(instance, [ 818.040021] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 818.040021] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] vif_infos = vmwarevif.get_vif_info(self._session, [ 818.040021] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 818.040021] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] for vif in network_info: [ 818.040021] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 818.040021] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] return self._sync_wrapper(fn, *args, **kwargs) [ 818.040021] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 818.040021] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] self.wait() [ 818.040021] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 818.040021] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] self[:] = self._gt.wait() [ 818.040021] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 818.040021] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] return self._exit_event.wait() [ 818.040021] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 818.040439] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] result = hub.switch() [ 818.040439] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 818.040439] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] return self.greenlet.switch() [ 818.040439] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 818.040439] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] result = function(*args, **kwargs) [ 818.040439] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 818.040439] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] return func(*args, **kwargs) [ 818.040439] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 818.040439] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] raise e [ 818.040439] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 818.040439] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] nwinfo = self.network_api.allocate_for_instance( [ 818.040439] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 818.040439] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] created_port_ids = self._update_ports_for_instance( [ 818.040870] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 818.040870] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] with excutils.save_and_reraise_exception(): [ 818.040870] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 818.040870] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] self.force_reraise() [ 818.040870] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 818.040870] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] raise self.value [ 818.040870] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 818.040870] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] updated_port = self._update_port( [ 818.040870] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 818.040870] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] _ensure_no_port_binding_failure(port) [ 818.040870] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 818.040870] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] raise exception.PortBindingFailed(port_id=port['id']) [ 818.041260] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] nova.exception.PortBindingFailed: Binding failed for port c5ae85f8-f1b8-497a-a7b9-a9b53a3c813c, please check neutron logs for more information. [ 818.041260] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] [ 818.041260] env[62952]: INFO nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Terminating instance [ 818.042985] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "refresh_cache-9c5167eb-3a15-47ff-923e-c83cc3529cc8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.227702] env[62952]: DEBUG nova.network.neutron [req-dd138084-a895-4c7c-a2ba-919f06fff5c0 req-344aff31-8270-4110-8d26-2877e6ba3115 service nova] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 818.344888] env[62952]: DEBUG nova.network.neutron [req-dd138084-a895-4c7c-a2ba-919f06fff5c0 req-344aff31-8270-4110-8d26-2877e6ba3115 service nova] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.366573] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.367231] env[62952]: DEBUG nova.compute.manager [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 818.369749] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.171s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.369989] env[62952]: DEBUG nova.objects.instance [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Lazy-loading 'resources' on Instance uuid 472e60f1-9b6c-415e-8a92-19235ad2e547 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 818.453679] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.847978] env[62952]: DEBUG oslo_concurrency.lockutils [req-dd138084-a895-4c7c-a2ba-919f06fff5c0 req-344aff31-8270-4110-8d26-2877e6ba3115 service nova] Releasing lock "refresh_cache-9c5167eb-3a15-47ff-923e-c83cc3529cc8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.848429] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquired lock "refresh_cache-9c5167eb-3a15-47ff-923e-c83cc3529cc8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.848617] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 818.875681] env[62952]: DEBUG nova.compute.utils [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 818.879168] env[62952]: DEBUG nova.compute.manager [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 818.879404] env[62952]: DEBUG nova.network.neutron [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 818.947348] env[62952]: DEBUG nova.policy [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'efe1248d9a294603b761ec7fede12d38', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7594174ba3e42e79dacd4ee9df575ad', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 819.156650] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132d05bb-7951-4f4a-b018-927eb20bc4b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.164695] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47023b7d-9a16-4166-bb47-be7a9988c6ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.199872] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cf0d481-0524-47cc-aee5-d771d605a110 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.207595] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eab48a2-eb87-46c0-b625-31c7b7df28e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.222324] env[62952]: DEBUG nova.compute.provider_tree [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.377959] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 819.385026] env[62952]: DEBUG nova.compute.manager [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 819.526209] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.676437] env[62952]: DEBUG nova.network.neutron [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Successfully created port: d1ec9d86-9392-41f9-95b2-2d574a3b9ef1 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 819.729612] env[62952]: DEBUG nova.scheduler.client.report [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 819.886452] env[62952]: DEBUG nova.compute.manager [req-0bd88eae-c4b9-4b80-8b03-c9d65ee66e22 req-c4fd0bc2-0502-4716-937f-006453886151 service nova] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Received event network-vif-deleted-c5ae85f8-f1b8-497a-a7b9-a9b53a3c813c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 820.028905] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Releasing lock "refresh_cache-9c5167eb-3a15-47ff-923e-c83cc3529cc8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.029350] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 820.029537] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 820.029912] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2aceeb0d-50d4-42ec-9339-1c056d63368f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.039045] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cc35ad3-ca7b-4e19-8931-811f23d87b36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.061196] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9c5167eb-3a15-47ff-923e-c83cc3529cc8 could not be found. [ 820.061650] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 820.061906] env[62952]: INFO nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Took 0.03 seconds to destroy the instance on the hypervisor. [ 820.062217] env[62952]: DEBUG oslo.service.loopingcall [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.062746] env[62952]: DEBUG nova.compute.manager [-] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 820.062889] env[62952]: DEBUG nova.network.neutron [-] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 820.085953] env[62952]: DEBUG nova.network.neutron [-] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 820.235832] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.866s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.240044] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.183s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.264943] env[62952]: INFO nova.scheduler.client.report [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Deleted allocations for instance 472e60f1-9b6c-415e-8a92-19235ad2e547 [ 820.394360] env[62952]: DEBUG nova.compute.manager [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 820.413509] env[62952]: DEBUG nova.virt.hardware [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 820.413766] env[62952]: DEBUG nova.virt.hardware [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 820.413923] env[62952]: DEBUG nova.virt.hardware [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 820.414115] env[62952]: DEBUG nova.virt.hardware [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 820.414260] env[62952]: DEBUG nova.virt.hardware [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 820.414404] env[62952]: DEBUG nova.virt.hardware [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 820.414610] env[62952]: DEBUG nova.virt.hardware [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 820.414768] env[62952]: DEBUG nova.virt.hardware [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 820.415241] env[62952]: DEBUG nova.virt.hardware [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 820.415494] env[62952]: DEBUG nova.virt.hardware [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 820.415866] env[62952]: DEBUG nova.virt.hardware [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 820.416726] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd39579-97d1-4ae4-ac00-33678c1187c4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.426706] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72206a91-3a7c-4a69-b8c8-794135e03163 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.588239] env[62952]: DEBUG nova.network.neutron [-] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.775819] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7703446a-4a68-4d0d-9640-438732b1d307 tempest-ServersAdmin275Test-653252852 tempest-ServersAdmin275Test-653252852-project-member] Lock "472e60f1-9b6c-415e-8a92-19235ad2e547" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.319s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.089205] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcd8a01e-be08-41b6-a7ef-72f07744e6cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.093108] env[62952]: INFO nova.compute.manager [-] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Took 1.03 seconds to deallocate network for instance. [ 821.096922] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c9ea39-a509-4450-9845-7adc5a75fabf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.102274] env[62952]: DEBUG nova.compute.claims [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 821.102274] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.134686] env[62952]: ERROR nova.compute.manager [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d1ec9d86-9392-41f9-95b2-2d574a3b9ef1, please check neutron logs for more information. [ 821.134686] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 821.134686] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 821.134686] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 821.134686] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 821.134686] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 821.134686] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 821.134686] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 821.134686] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 821.134686] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 821.134686] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 821.134686] env[62952]: ERROR nova.compute.manager raise self.value [ 821.134686] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 821.134686] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 821.134686] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 821.134686] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 821.135190] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 821.135190] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 821.135190] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d1ec9d86-9392-41f9-95b2-2d574a3b9ef1, please check neutron logs for more information. [ 821.135190] env[62952]: ERROR nova.compute.manager [ 821.139019] env[62952]: Traceback (most recent call last): [ 821.139019] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 821.139019] env[62952]: listener.cb(fileno) [ 821.139019] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 821.139019] env[62952]: result = function(*args, **kwargs) [ 821.139019] env[62952]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 821.139019] env[62952]: return func(*args, **kwargs) [ 821.139019] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 821.139019] env[62952]: raise e [ 821.139019] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 821.139019] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 821.139019] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 821.139019] env[62952]: created_port_ids = self._update_ports_for_instance( [ 821.139019] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 821.139019] env[62952]: with excutils.save_and_reraise_exception(): [ 821.139019] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 821.139019] env[62952]: self.force_reraise() [ 821.139019] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 821.139019] env[62952]: raise self.value [ 821.139019] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 821.139019] env[62952]: updated_port = self._update_port( [ 821.139019] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 821.139019] env[62952]: _ensure_no_port_binding_failure(port) [ 821.139019] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 821.139019] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 821.139019] env[62952]: nova.exception.PortBindingFailed: Binding failed for port d1ec9d86-9392-41f9-95b2-2d574a3b9ef1, please check neutron logs for more information. [ 821.139019] env[62952]: Removing descriptor: 19 [ 821.139019] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6063c64-1ec2-497a-bd8f-53ad8833be52 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.140520] env[62952]: ERROR nova.compute.manager [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d1ec9d86-9392-41f9-95b2-2d574a3b9ef1, please check neutron logs for more information. [ 821.140520] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Traceback (most recent call last): [ 821.140520] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 821.140520] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] yield resources [ 821.140520] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 821.140520] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] self.driver.spawn(context, instance, image_meta, [ 821.140520] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 821.140520] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] self._vmops.spawn(context, instance, image_meta, injected_files, [ 821.140520] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 821.140520] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] vm_ref = self.build_virtual_machine(instance, [ 821.140520] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 821.140939] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] vif_infos = vmwarevif.get_vif_info(self._session, [ 821.140939] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 821.140939] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] for vif in network_info: [ 821.140939] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 821.140939] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] return self._sync_wrapper(fn, *args, **kwargs) [ 821.140939] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 821.140939] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] self.wait() [ 821.140939] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 821.140939] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] self[:] = self._gt.wait() [ 821.140939] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 821.140939] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] return self._exit_event.wait() [ 821.140939] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 821.140939] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] result = hub.switch() [ 821.141336] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 821.141336] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] return self.greenlet.switch() [ 821.141336] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 821.141336] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] result = function(*args, **kwargs) [ 821.141336] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 821.141336] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] return func(*args, **kwargs) [ 821.141336] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 821.141336] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] raise e [ 821.141336] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 821.141336] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] nwinfo = self.network_api.allocate_for_instance( [ 821.141336] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 821.141336] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] created_port_ids = self._update_ports_for_instance( [ 821.141336] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 821.141716] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] with excutils.save_and_reraise_exception(): [ 821.141716] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 821.141716] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] self.force_reraise() [ 821.141716] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 821.141716] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] raise self.value [ 821.141716] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 821.141716] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] updated_port = self._update_port( [ 821.141716] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 821.141716] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] _ensure_no_port_binding_failure(port) [ 821.141716] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 821.141716] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] raise exception.PortBindingFailed(port_id=port['id']) [ 821.141716] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] nova.exception.PortBindingFailed: Binding failed for port d1ec9d86-9392-41f9-95b2-2d574a3b9ef1, please check neutron logs for more information. [ 821.141716] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] [ 821.142121] env[62952]: INFO nova.compute.manager [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Terminating instance [ 821.143691] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Acquiring lock "refresh_cache-9be8f7f5-80eb-4e38-9501-e76abd7227db" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.143967] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Acquired lock "refresh_cache-9be8f7f5-80eb-4e38-9501-e76abd7227db" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.145605] env[62952]: DEBUG nova.network.neutron [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 821.149016] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ad1a4e-22fb-4361-b831-d3aef7f4491d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.163952] env[62952]: DEBUG nova.compute.provider_tree [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.667400] env[62952]: DEBUG nova.network.neutron [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 821.673376] env[62952]: DEBUG nova.scheduler.client.report [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 821.767895] env[62952]: DEBUG nova.network.neutron [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.927612] env[62952]: DEBUG nova.compute.manager [req-9915cfb1-d015-4910-b4ba-f8ee253fd5e0 req-696f63c7-bc63-45cd-8053-5c4589f0b28f service nova] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Received event network-changed-d1ec9d86-9392-41f9-95b2-2d574a3b9ef1 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 821.927888] env[62952]: DEBUG nova.compute.manager [req-9915cfb1-d015-4910-b4ba-f8ee253fd5e0 req-696f63c7-bc63-45cd-8053-5c4589f0b28f service nova] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Refreshing instance network info cache due to event network-changed-d1ec9d86-9392-41f9-95b2-2d574a3b9ef1. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 821.928119] env[62952]: DEBUG oslo_concurrency.lockutils [req-9915cfb1-d015-4910-b4ba-f8ee253fd5e0 req-696f63c7-bc63-45cd-8053-5c4589f0b28f service nova] Acquiring lock "refresh_cache-9be8f7f5-80eb-4e38-9501-e76abd7227db" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.178695] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.938s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.179362] env[62952]: ERROR nova.compute.manager [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6629a780-8685-4991-9015-0601155c4978, please check neutron logs for more information. [ 822.179362] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Traceback (most recent call last): [ 822.179362] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 822.179362] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] self.driver.spawn(context, instance, image_meta, [ 822.179362] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 822.179362] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 822.179362] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 822.179362] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] vm_ref = self.build_virtual_machine(instance, [ 822.179362] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 822.179362] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] vif_infos = vmwarevif.get_vif_info(self._session, [ 822.179362] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 822.179786] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] for vif in network_info: [ 822.179786] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 822.179786] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] return self._sync_wrapper(fn, *args, **kwargs) [ 822.179786] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 822.179786] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] self.wait() [ 822.179786] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 822.179786] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] self[:] = self._gt.wait() [ 822.179786] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 822.179786] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] return self._exit_event.wait() [ 822.179786] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 822.179786] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] current.throw(*self._exc) [ 822.179786] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 822.179786] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] result = function(*args, **kwargs) [ 822.180246] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 822.180246] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] return func(*args, **kwargs) [ 822.180246] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 822.180246] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] raise e [ 822.180246] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 822.180246] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] nwinfo = self.network_api.allocate_for_instance( [ 822.180246] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 822.180246] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] created_port_ids = self._update_ports_for_instance( [ 822.180246] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 822.180246] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] with excutils.save_and_reraise_exception(): [ 822.180246] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 822.180246] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] self.force_reraise() [ 822.180246] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 822.180686] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] raise self.value [ 822.180686] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 822.180686] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] updated_port = self._update_port( [ 822.180686] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 822.180686] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] _ensure_no_port_binding_failure(port) [ 822.180686] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 822.180686] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] raise exception.PortBindingFailed(port_id=port['id']) [ 822.180686] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] nova.exception.PortBindingFailed: Binding failed for port 6629a780-8685-4991-9015-0601155c4978, please check neutron logs for more information. [ 822.180686] env[62952]: ERROR nova.compute.manager [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] [ 822.180686] env[62952]: DEBUG nova.compute.utils [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Binding failed for port 6629a780-8685-4991-9015-0601155c4978, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 822.181138] env[62952]: DEBUG oslo_concurrency.lockutils [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.338s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.182568] env[62952]: INFO nova.compute.claims [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 822.185194] env[62952]: DEBUG nova.compute.manager [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Build of instance ebff0c87-8e58-4a76-a350-43a3d660debd was re-scheduled: Binding failed for port 6629a780-8685-4991-9015-0601155c4978, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 822.185560] env[62952]: DEBUG nova.compute.manager [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 822.185762] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Acquiring lock "refresh_cache-ebff0c87-8e58-4a76-a350-43a3d660debd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.185921] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Acquired lock "refresh_cache-ebff0c87-8e58-4a76-a350-43a3d660debd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.186087] env[62952]: DEBUG nova.network.neutron [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 822.273616] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Releasing lock "refresh_cache-9be8f7f5-80eb-4e38-9501-e76abd7227db" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.274097] env[62952]: DEBUG nova.compute.manager [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 822.274263] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 822.274653] env[62952]: DEBUG oslo_concurrency.lockutils [req-9915cfb1-d015-4910-b4ba-f8ee253fd5e0 req-696f63c7-bc63-45cd-8053-5c4589f0b28f service nova] Acquired lock "refresh_cache-9be8f7f5-80eb-4e38-9501-e76abd7227db" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.274740] env[62952]: DEBUG nova.network.neutron [req-9915cfb1-d015-4910-b4ba-f8ee253fd5e0 req-696f63c7-bc63-45cd-8053-5c4589f0b28f service nova] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Refreshing network info cache for port d1ec9d86-9392-41f9-95b2-2d574a3b9ef1 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 822.276239] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f6187bda-3ab9-4814-bfff-3eb98ea08bbe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.285876] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6699e91a-da4d-4cb8-b471-28bc174c7986 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.307985] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9be8f7f5-80eb-4e38-9501-e76abd7227db could not be found. [ 822.308496] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 822.308496] env[62952]: INFO nova.compute.manager [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Took 0.03 seconds to destroy the instance on the hypervisor. [ 822.308701] env[62952]: DEBUG oslo.service.loopingcall [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.309397] env[62952]: DEBUG nova.compute.manager [-] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 822.309397] env[62952]: DEBUG nova.network.neutron [-] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 822.331392] env[62952]: DEBUG nova.network.neutron [-] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 822.486772] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Acquiring lock "49947bde-069b-459c-ae19-ae5f090f535b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.486772] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Lock "49947bde-069b-459c-ae19-ae5f090f535b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.708857] env[62952]: DEBUG nova.network.neutron [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 822.810125] env[62952]: DEBUG nova.network.neutron [req-9915cfb1-d015-4910-b4ba-f8ee253fd5e0 req-696f63c7-bc63-45cd-8053-5c4589f0b28f service nova] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 822.833735] env[62952]: DEBUG nova.network.neutron [-] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.871975] env[62952]: DEBUG nova.network.neutron [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.935475] env[62952]: DEBUG nova.network.neutron [req-9915cfb1-d015-4910-b4ba-f8ee253fd5e0 req-696f63c7-bc63-45cd-8053-5c4589f0b28f service nova] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.336162] env[62952]: INFO nova.compute.manager [-] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Took 1.03 seconds to deallocate network for instance. [ 823.338761] env[62952]: DEBUG nova.compute.claims [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 823.338891] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.376483] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Releasing lock "refresh_cache-ebff0c87-8e58-4a76-a350-43a3d660debd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.376483] env[62952]: DEBUG nova.compute.manager [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 823.376483] env[62952]: DEBUG nova.compute.manager [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 823.376483] env[62952]: DEBUG nova.network.neutron [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 823.397090] env[62952]: DEBUG nova.network.neutron [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 823.439229] env[62952]: DEBUG oslo_concurrency.lockutils [req-9915cfb1-d015-4910-b4ba-f8ee253fd5e0 req-696f63c7-bc63-45cd-8053-5c4589f0b28f service nova] Releasing lock "refresh_cache-9be8f7f5-80eb-4e38-9501-e76abd7227db" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.439229] env[62952]: DEBUG nova.compute.manager [req-9915cfb1-d015-4910-b4ba-f8ee253fd5e0 req-696f63c7-bc63-45cd-8053-5c4589f0b28f service nova] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Received event network-vif-deleted-d1ec9d86-9392-41f9-95b2-2d574a3b9ef1 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 823.474018] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3cb6147-2e34-47b3-8ea0-60ffcd0f84c4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.481722] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac59acc-ee08-4af6-a729-2c9235bf98b1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.515557] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e8f603-d1e5-4176-a0ab-3c88f5e9e237 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.525165] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dcd7c10-ecbc-4f4d-b3c4-db21e14acc0b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.540837] env[62952]: DEBUG nova.compute.provider_tree [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 823.898929] env[62952]: DEBUG nova.network.neutron [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.075682] env[62952]: ERROR nova.scheduler.client.report [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [req-7d578bd5-dc14-45a7-b093-9f1338e5fcd8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 615127b5-dc82-4491-9c8f-4e6d0caa0690. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7d578bd5-dc14-45a7-b093-9f1338e5fcd8"}]} [ 824.101028] env[62952]: DEBUG nova.scheduler.client.report [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Refreshing inventories for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 824.128352] env[62952]: DEBUG nova.scheduler.client.report [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Updating ProviderTree inventory for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 824.128564] env[62952]: DEBUG nova.compute.provider_tree [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 824.145156] env[62952]: DEBUG nova.scheduler.client.report [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Refreshing aggregate associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 824.167557] env[62952]: DEBUG nova.scheduler.client.report [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Refreshing trait associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 824.401400] env[62952]: INFO nova.compute.manager [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] [instance: ebff0c87-8e58-4a76-a350-43a3d660debd] Took 1.03 seconds to deallocate network for instance. [ 824.505961] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7075721-6041-4dd9-89c7-e7658fd6b370 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.515103] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee5e9fbb-82cf-4b28-89d5-18b7b11c6075 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.548878] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec8e084-61f0-4186-b266-b405a1d13a2b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.557979] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e166801b-f7e1-4642-8a6b-4fa5127dde8e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.572926] env[62952]: DEBUG nova.compute.provider_tree [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 825.094688] env[62952]: ERROR nova.scheduler.client.report [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [req-3cbd1e94-d943-4f69-b157-67524e75e7f9] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 615127b5-dc82-4491-9c8f-4e6d0caa0690. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3cbd1e94-d943-4f69-b157-67524e75e7f9"}]} [ 825.115590] env[62952]: DEBUG nova.scheduler.client.report [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Refreshing inventories for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 825.132432] env[62952]: DEBUG nova.scheduler.client.report [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Updating ProviderTree inventory for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 825.132655] env[62952]: DEBUG nova.compute.provider_tree [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 825.147682] env[62952]: DEBUG nova.scheduler.client.report [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Refreshing aggregate associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 825.167019] env[62952]: DEBUG nova.scheduler.client.report [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Refreshing trait associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 825.431662] env[62952]: INFO nova.scheduler.client.report [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Deleted allocations for instance ebff0c87-8e58-4a76-a350-43a3d660debd [ 825.450792] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "ba022bea-e357-4541-a573-d1da2eb49b46" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.450792] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "ba022bea-e357-4541-a573-d1da2eb49b46" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.452566] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c0cb98-047e-4b54-8e82-1fb0fcb95f0d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.461670] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64b5e46c-0888-4b7a-8816-6e9a8c034718 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.492408] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6bfb94f-08b8-47b6-9c3a-e276a51f17c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.499870] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26007d7-c348-4808-99f9-97a1ad080063 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.513242] env[62952]: DEBUG nova.compute.provider_tree [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 825.942448] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8a61253b-fa5c-4767-86d4-317aaa696204 tempest-ServersNegativeTestJSON-803059911 tempest-ServersNegativeTestJSON-803059911-project-member] Lock "ebff0c87-8e58-4a76-a350-43a3d660debd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.071s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.063178] env[62952]: DEBUG nova.scheduler.client.report [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Updated inventory for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with generation 75 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 826.063178] env[62952]: DEBUG nova.compute.provider_tree [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Updating resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 generation from 75 to 76 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 826.063178] env[62952]: DEBUG nova.compute.provider_tree [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 826.448817] env[62952]: DEBUG nova.compute.manager [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 826.571019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.388s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.571019] env[62952]: DEBUG nova.compute.manager [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 826.573098] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.851s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.575180] env[62952]: INFO nova.compute.claims [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.978507] env[62952]: DEBUG oslo_concurrency.lockutils [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.079871] env[62952]: DEBUG nova.compute.utils [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 827.083463] env[62952]: DEBUG nova.compute.manager [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 827.083687] env[62952]: DEBUG nova.network.neutron [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 827.146886] env[62952]: DEBUG nova.policy [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a2a6924fed644c8fb8efc38f735df7af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b8ad573f9d144a0ba78e1b666fd5660', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 827.529467] env[62952]: DEBUG nova.network.neutron [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Successfully created port: 6af48a64-0373-41eb-a15f-976266b85305 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 827.587734] env[62952]: DEBUG nova.compute.manager [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 827.915019] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46689d1d-1818-4f56-b777-072056ffb0b3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.923793] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b958036a-ce0d-4e24-9da6-8b66c78622cd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.962082] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f4195b-e516-4a65-b8d0-619a02969c1e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.969253] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514c9edd-ec2a-4fc5-8c4d-7a0cdce72623 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.984462] env[62952]: DEBUG nova.compute.provider_tree [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.488742] env[62952]: DEBUG nova.scheduler.client.report [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.598959] env[62952]: DEBUG nova.compute.manager [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 828.620142] env[62952]: DEBUG nova.virt.hardware [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.620418] env[62952]: DEBUG nova.virt.hardware [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.620577] env[62952]: DEBUG nova.virt.hardware [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.620765] env[62952]: DEBUG nova.virt.hardware [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.620912] env[62952]: DEBUG nova.virt.hardware [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.621081] env[62952]: DEBUG nova.virt.hardware [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.621333] env[62952]: DEBUG nova.virt.hardware [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.621494] env[62952]: DEBUG nova.virt.hardware [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.621656] env[62952]: DEBUG nova.virt.hardware [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.621814] env[62952]: DEBUG nova.virt.hardware [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.621981] env[62952]: DEBUG nova.virt.hardware [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.622966] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b592cc0-7be5-4551-a619-7377782ec6b6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.630906] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4008e983-edaf-4f81-afde-10ce17d5a053 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.994952] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.994952] env[62952]: DEBUG nova.compute.manager [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 828.997323] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.249s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.115668] env[62952]: DEBUG nova.compute.manager [req-f236695f-9e2d-4a69-ae5b-dee98c9ba994 req-694c56d6-5353-4b96-be29-8571d1fc69f9 service nova] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Received event network-vif-plugged-6af48a64-0373-41eb-a15f-976266b85305 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 829.115668] env[62952]: DEBUG oslo_concurrency.lockutils [req-f236695f-9e2d-4a69-ae5b-dee98c9ba994 req-694c56d6-5353-4b96-be29-8571d1fc69f9 service nova] Acquiring lock "8ac98d96-139a-4a03-94f3-178fafa88503-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.116112] env[62952]: DEBUG oslo_concurrency.lockutils [req-f236695f-9e2d-4a69-ae5b-dee98c9ba994 req-694c56d6-5353-4b96-be29-8571d1fc69f9 service nova] Lock "8ac98d96-139a-4a03-94f3-178fafa88503-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.116112] env[62952]: DEBUG oslo_concurrency.lockutils [req-f236695f-9e2d-4a69-ae5b-dee98c9ba994 req-694c56d6-5353-4b96-be29-8571d1fc69f9 service nova] Lock "8ac98d96-139a-4a03-94f3-178fafa88503-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.116333] env[62952]: DEBUG nova.compute.manager [req-f236695f-9e2d-4a69-ae5b-dee98c9ba994 req-694c56d6-5353-4b96-be29-8571d1fc69f9 service nova] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] No waiting events found dispatching network-vif-plugged-6af48a64-0373-41eb-a15f-976266b85305 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 829.116404] env[62952]: WARNING nova.compute.manager [req-f236695f-9e2d-4a69-ae5b-dee98c9ba994 req-694c56d6-5353-4b96-be29-8571d1fc69f9 service nova] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Received unexpected event network-vif-plugged-6af48a64-0373-41eb-a15f-976266b85305 for instance with vm_state building and task_state spawning. [ 829.134998] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquiring lock "14f8dbbe-2c4a-4948-bf15-106d5b1b0677" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.134998] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "14f8dbbe-2c4a-4948-bf15-106d5b1b0677" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.295723] env[62952]: DEBUG oslo_concurrency.lockutils [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquiring lock "48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.295723] env[62952]: DEBUG oslo_concurrency.lockutils [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.501782] env[62952]: DEBUG nova.compute.utils [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.506965] env[62952]: DEBUG nova.compute.manager [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 829.507181] env[62952]: DEBUG nova.network.neutron [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 829.558513] env[62952]: DEBUG nova.network.neutron [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Successfully updated port: 6af48a64-0373-41eb-a15f-976266b85305 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 829.567736] env[62952]: DEBUG nova.policy [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e872d861b56e435d8c1250dbb82a5dfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8960909bb80f4395bb87b6de3c60ab2c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 829.582938] env[62952]: DEBUG nova.compute.manager [req-3feffc2c-2a09-4c9d-9efc-3d8eab794875 req-d494e61c-726d-4bea-96b7-ea16d195b8aa service nova] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Received event network-changed-6af48a64-0373-41eb-a15f-976266b85305 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 829.583146] env[62952]: DEBUG nova.compute.manager [req-3feffc2c-2a09-4c9d-9efc-3d8eab794875 req-d494e61c-726d-4bea-96b7-ea16d195b8aa service nova] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Refreshing instance network info cache due to event network-changed-6af48a64-0373-41eb-a15f-976266b85305. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 829.583364] env[62952]: DEBUG oslo_concurrency.lockutils [req-3feffc2c-2a09-4c9d-9efc-3d8eab794875 req-d494e61c-726d-4bea-96b7-ea16d195b8aa service nova] Acquiring lock "refresh_cache-8ac98d96-139a-4a03-94f3-178fafa88503" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.583500] env[62952]: DEBUG oslo_concurrency.lockutils [req-3feffc2c-2a09-4c9d-9efc-3d8eab794875 req-d494e61c-726d-4bea-96b7-ea16d195b8aa service nova] Acquired lock "refresh_cache-8ac98d96-139a-4a03-94f3-178fafa88503" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.583647] env[62952]: DEBUG nova.network.neutron [req-3feffc2c-2a09-4c9d-9efc-3d8eab794875 req-d494e61c-726d-4bea-96b7-ea16d195b8aa service nova] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Refreshing network info cache for port 6af48a64-0373-41eb-a15f-976266b85305 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 829.823097] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c1f5e1-baac-4e63-bf4d-1ca44ff88493 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.831996] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3479d97c-9b37-48d9-bcab-ad3322ad4d7d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.866463] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72cae2b3-f69f-4695-990d-d3fffd4981f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.874085] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dca42c6-fb93-4821-9457-1414333813ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.887605] env[62952]: DEBUG nova.compute.provider_tree [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.889560] env[62952]: DEBUG nova.network.neutron [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Successfully created port: 363ca83f-c6ac-4446-996c-edc459eb29bf {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 830.010638] env[62952]: DEBUG nova.compute.manager [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 830.061031] env[62952]: DEBUG oslo_concurrency.lockutils [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Acquiring lock "refresh_cache-8ac98d96-139a-4a03-94f3-178fafa88503" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.132252] env[62952]: DEBUG nova.network.neutron [req-3feffc2c-2a09-4c9d-9efc-3d8eab794875 req-d494e61c-726d-4bea-96b7-ea16d195b8aa service nova] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 830.218468] env[62952]: DEBUG nova.network.neutron [req-3feffc2c-2a09-4c9d-9efc-3d8eab794875 req-d494e61c-726d-4bea-96b7-ea16d195b8aa service nova] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.392857] env[62952]: DEBUG nova.scheduler.client.report [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 830.722267] env[62952]: DEBUG oslo_concurrency.lockutils [req-3feffc2c-2a09-4c9d-9efc-3d8eab794875 req-d494e61c-726d-4bea-96b7-ea16d195b8aa service nova] Releasing lock "refresh_cache-8ac98d96-139a-4a03-94f3-178fafa88503" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.722703] env[62952]: DEBUG oslo_concurrency.lockutils [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Acquired lock "refresh_cache-8ac98d96-139a-4a03-94f3-178fafa88503" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.722876] env[62952]: DEBUG nova.network.neutron [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 830.897577] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.900s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.898358] env[62952]: ERROR nova.compute.manager [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f55e5aac-9df2-468a-8f4c-13216c743efc, please check neutron logs for more information. [ 830.898358] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Traceback (most recent call last): [ 830.898358] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 830.898358] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] self.driver.spawn(context, instance, image_meta, [ 830.898358] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 830.898358] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 830.898358] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 830.898358] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] vm_ref = self.build_virtual_machine(instance, [ 830.898358] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 830.898358] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] vif_infos = vmwarevif.get_vif_info(self._session, [ 830.898358] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 830.898812] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] for vif in network_info: [ 830.898812] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 830.898812] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] return self._sync_wrapper(fn, *args, **kwargs) [ 830.898812] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 830.898812] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] self.wait() [ 830.898812] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 830.898812] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] self[:] = self._gt.wait() [ 830.898812] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 830.898812] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] return self._exit_event.wait() [ 830.898812] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 830.898812] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] result = hub.switch() [ 830.898812] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 830.898812] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] return self.greenlet.switch() [ 830.899215] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 830.899215] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] result = function(*args, **kwargs) [ 830.899215] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 830.899215] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] return func(*args, **kwargs) [ 830.899215] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 830.899215] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] raise e [ 830.899215] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 830.899215] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] nwinfo = self.network_api.allocate_for_instance( [ 830.899215] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 830.899215] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] created_port_ids = self._update_ports_for_instance( [ 830.899215] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 830.899215] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] with excutils.save_and_reraise_exception(): [ 830.899215] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 830.899753] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] self.force_reraise() [ 830.899753] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 830.899753] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] raise self.value [ 830.899753] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 830.899753] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] updated_port = self._update_port( [ 830.899753] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 830.899753] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] _ensure_no_port_binding_failure(port) [ 830.899753] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 830.899753] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] raise exception.PortBindingFailed(port_id=port['id']) [ 830.899753] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] nova.exception.PortBindingFailed: Binding failed for port f55e5aac-9df2-468a-8f4c-13216c743efc, please check neutron logs for more information. [ 830.899753] env[62952]: ERROR nova.compute.manager [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] [ 830.900242] env[62952]: DEBUG nova.compute.utils [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Binding failed for port f55e5aac-9df2-468a-8f4c-13216c743efc, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 830.900355] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.781s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.902114] env[62952]: INFO nova.compute.claims [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 830.906021] env[62952]: DEBUG nova.compute.manager [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Build of instance 1ed9de84-6178-4173-84ad-dc115be7a1fa was re-scheduled: Binding failed for port f55e5aac-9df2-468a-8f4c-13216c743efc, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 830.906021] env[62952]: DEBUG nova.compute.manager [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 830.906021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Acquiring lock "refresh_cache-1ed9de84-6178-4173-84ad-dc115be7a1fa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.906021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Acquired lock "refresh_cache-1ed9de84-6178-4173-84ad-dc115be7a1fa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.906246] env[62952]: DEBUG nova.network.neutron [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 831.021238] env[62952]: DEBUG nova.compute.manager [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 831.045479] env[62952]: DEBUG nova.virt.hardware [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 831.045752] env[62952]: DEBUG nova.virt.hardware [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 831.045908] env[62952]: DEBUG nova.virt.hardware [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.046096] env[62952]: DEBUG nova.virt.hardware [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 831.046245] env[62952]: DEBUG nova.virt.hardware [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.046392] env[62952]: DEBUG nova.virt.hardware [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 831.046594] env[62952]: DEBUG nova.virt.hardware [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 831.046749] env[62952]: DEBUG nova.virt.hardware [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 831.046913] env[62952]: DEBUG nova.virt.hardware [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 831.047089] env[62952]: DEBUG nova.virt.hardware [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 831.047265] env[62952]: DEBUG nova.virt.hardware [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 831.048141] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b200330-14da-4bf4-9cdd-5d2328e875d6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.055720] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30daf4da-b420-4c8c-9873-65062c42a5af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.272320] env[62952]: DEBUG nova.network.neutron [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 831.453631] env[62952]: DEBUG nova.network.neutron [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 831.594585] env[62952]: DEBUG nova.compute.manager [req-2b6343df-c1bf-42fd-a83a-66a06b8cd6d3 req-0fa86e34-7a57-46bc-9d9e-b0c3cc90e44b service nova] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Received event network-vif-plugged-363ca83f-c6ac-4446-996c-edc459eb29bf {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 831.594585] env[62952]: DEBUG oslo_concurrency.lockutils [req-2b6343df-c1bf-42fd-a83a-66a06b8cd6d3 req-0fa86e34-7a57-46bc-9d9e-b0c3cc90e44b service nova] Acquiring lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.594585] env[62952]: DEBUG oslo_concurrency.lockutils [req-2b6343df-c1bf-42fd-a83a-66a06b8cd6d3 req-0fa86e34-7a57-46bc-9d9e-b0c3cc90e44b service nova] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.594585] env[62952]: DEBUG oslo_concurrency.lockutils [req-2b6343df-c1bf-42fd-a83a-66a06b8cd6d3 req-0fa86e34-7a57-46bc-9d9e-b0c3cc90e44b service nova] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.594926] env[62952]: DEBUG nova.compute.manager [req-2b6343df-c1bf-42fd-a83a-66a06b8cd6d3 req-0fa86e34-7a57-46bc-9d9e-b0c3cc90e44b service nova] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] No waiting events found dispatching network-vif-plugged-363ca83f-c6ac-4446-996c-edc459eb29bf {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 831.594926] env[62952]: WARNING nova.compute.manager [req-2b6343df-c1bf-42fd-a83a-66a06b8cd6d3 req-0fa86e34-7a57-46bc-9d9e-b0c3cc90e44b service nova] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Received unexpected event network-vif-plugged-363ca83f-c6ac-4446-996c-edc459eb29bf for instance with vm_state building and task_state spawning. [ 831.629378] env[62952]: DEBUG nova.network.neutron [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.677587] env[62952]: DEBUG nova.network.neutron [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Updating instance_info_cache with network_info: [{"id": "6af48a64-0373-41eb-a15f-976266b85305", "address": "fa:16:3e:c9:25:55", "network": {"id": "7e018521-7c32-4e53-b20d-6b1b0872c723", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1074718007-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b8ad573f9d144a0ba78e1b666fd5660", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "23fc30ea-1f06-424d-86e1-27ae5435b1a9", "external-id": "nsx-vlan-transportzone-189", "segmentation_id": 189, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6af48a64-03", "ovs_interfaceid": "6af48a64-0373-41eb-a15f-976266b85305", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.131491] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Releasing lock "refresh_cache-1ed9de84-6178-4173-84ad-dc115be7a1fa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.132516] env[62952]: DEBUG nova.compute.manager [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 832.132516] env[62952]: DEBUG nova.compute.manager [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 832.132516] env[62952]: DEBUG nova.network.neutron [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 832.158299] env[62952]: DEBUG nova.network.neutron [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 832.173904] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db3ba1a-042a-44a1-84e0-eedc00907570 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.177684] env[62952]: DEBUG nova.network.neutron [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Successfully updated port: 363ca83f-c6ac-4446-996c-edc459eb29bf {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 832.179554] env[62952]: DEBUG oslo_concurrency.lockutils [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Releasing lock "refresh_cache-8ac98d96-139a-4a03-94f3-178fafa88503" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.181257] env[62952]: DEBUG nova.compute.manager [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Instance network_info: |[{"id": "6af48a64-0373-41eb-a15f-976266b85305", "address": "fa:16:3e:c9:25:55", "network": {"id": "7e018521-7c32-4e53-b20d-6b1b0872c723", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1074718007-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b8ad573f9d144a0ba78e1b666fd5660", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "23fc30ea-1f06-424d-86e1-27ae5435b1a9", "external-id": "nsx-vlan-transportzone-189", "segmentation_id": 189, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6af48a64-03", "ovs_interfaceid": "6af48a64-0373-41eb-a15f-976266b85305", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 832.181520] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:25:55', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '23fc30ea-1f06-424d-86e1-27ae5435b1a9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6af48a64-0373-41eb-a15f-976266b85305', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 832.190745] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Creating folder: Project (0b8ad573f9d144a0ba78e1b666fd5660). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 832.194253] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-49f749b5-b237-4a17-ba5e-28e9acd5f4e0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.197394] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17320d90-b662-4bb3-90b0-53fbf079fb8c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.203453] env[62952]: DEBUG nova.compute.manager [req-4904b5a2-7916-4a1f-bfa7-2537db5709ed req-80d3c730-9b47-4db4-8a2b-2babaa97e928 service nova] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Received event network-changed-363ca83f-c6ac-4446-996c-edc459eb29bf {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 832.203637] env[62952]: DEBUG nova.compute.manager [req-4904b5a2-7916-4a1f-bfa7-2537db5709ed req-80d3c730-9b47-4db4-8a2b-2babaa97e928 service nova] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Refreshing instance network info cache due to event network-changed-363ca83f-c6ac-4446-996c-edc459eb29bf. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 832.203910] env[62952]: DEBUG oslo_concurrency.lockutils [req-4904b5a2-7916-4a1f-bfa7-2537db5709ed req-80d3c730-9b47-4db4-8a2b-2babaa97e928 service nova] Acquiring lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.204070] env[62952]: DEBUG oslo_concurrency.lockutils [req-4904b5a2-7916-4a1f-bfa7-2537db5709ed req-80d3c730-9b47-4db4-8a2b-2babaa97e928 service nova] Acquired lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.204281] env[62952]: DEBUG nova.network.neutron [req-4904b5a2-7916-4a1f-bfa7-2537db5709ed req-80d3c730-9b47-4db4-8a2b-2babaa97e928 service nova] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Refreshing network info cache for port 363ca83f-c6ac-4446-996c-edc459eb29bf {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 832.238814] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2edef79-9840-4553-b86d-ac42f6d23bb8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.241800] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Created folder: Project (0b8ad573f9d144a0ba78e1b666fd5660) in parent group-v271811. [ 832.241978] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Creating folder: Instances. Parent ref: group-v271831. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 832.242242] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0debe308-8b0f-4911-8e62-79e532550621 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.248917] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-084bfaf0-66d3-458c-9b46-adb8868667f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.253708] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Created folder: Instances in parent group-v271831. [ 832.253936] env[62952]: DEBUG oslo.service.loopingcall [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.254411] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 832.254613] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34b8f9e3-26cc-4e9f-a05a-8f5873f8bac1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.276871] env[62952]: DEBUG nova.compute.provider_tree [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.284240] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 832.284240] env[62952]: value = "task-1263132" [ 832.284240] env[62952]: _type = "Task" [ 832.284240] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.292067] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263132, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.660848] env[62952]: DEBUG nova.network.neutron [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.680495] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.737404] env[62952]: DEBUG nova.network.neutron [req-4904b5a2-7916-4a1f-bfa7-2537db5709ed req-80d3c730-9b47-4db4-8a2b-2babaa97e928 service nova] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 832.780831] env[62952]: DEBUG nova.scheduler.client.report [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 832.793017] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263132, 'name': CreateVM_Task, 'duration_secs': 0.298099} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.793217] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 832.805201] env[62952]: DEBUG oslo_concurrency.lockutils [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.805446] env[62952]: DEBUG oslo_concurrency.lockutils [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.805764] env[62952]: DEBUG oslo_concurrency.lockutils [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 832.806025] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92cb3100-0b61-4a35-b7f8-958a72817e6e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.816017] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Waiting for the task: (returnval){ [ 832.816017] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526509dd-117b-13d4-acae-bbf89ca14d34" [ 832.816017] env[62952]: _type = "Task" [ 832.816017] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.820911] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526509dd-117b-13d4-acae-bbf89ca14d34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.821727] env[62952]: DEBUG nova.network.neutron [req-4904b5a2-7916-4a1f-bfa7-2537db5709ed req-80d3c730-9b47-4db4-8a2b-2babaa97e928 service nova] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.165667] env[62952]: INFO nova.compute.manager [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] [instance: 1ed9de84-6178-4173-84ad-dc115be7a1fa] Took 1.03 seconds to deallocate network for instance. [ 833.285567] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.286088] env[62952]: DEBUG nova.compute.manager [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 833.288549] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.484s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.323262] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526509dd-117b-13d4-acae-bbf89ca14d34, 'name': SearchDatastore_Task, 'duration_secs': 0.010163} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.324369] env[62952]: DEBUG oslo_concurrency.lockutils [req-4904b5a2-7916-4a1f-bfa7-2537db5709ed req-80d3c730-9b47-4db4-8a2b-2babaa97e928 service nova] Releasing lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.324887] env[62952]: DEBUG oslo_concurrency.lockutils [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.325012] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 833.325256] env[62952]: DEBUG oslo_concurrency.lockutils [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.325400] env[62952]: DEBUG oslo_concurrency.lockutils [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.325577] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 833.325835] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquired lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.325978] env[62952]: DEBUG nova.network.neutron [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 833.327116] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-950dbbf4-1952-4e83-9b47-ef1fd8968d9f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.336312] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 833.336527] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 833.337265] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b863ad3-2a54-4e6b-8eda-8e5c0138c708 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.342388] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Waiting for the task: (returnval){ [ 833.342388] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52cb95c0-7eca-4dcf-1dd2-59257412b6ed" [ 833.342388] env[62952]: _type = "Task" [ 833.342388] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.350378] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52cb95c0-7eca-4dcf-1dd2-59257412b6ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.794364] env[62952]: DEBUG nova.compute.utils [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 833.799168] env[62952]: DEBUG nova.compute.manager [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 833.799168] env[62952]: DEBUG nova.network.neutron [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 833.839293] env[62952]: DEBUG nova.policy [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9f6f0133f4a4ac287d4d8d3f4992754', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1487c06883e444a589d9b688155dc00e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 833.852905] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52cb95c0-7eca-4dcf-1dd2-59257412b6ed, 'name': SearchDatastore_Task, 'duration_secs': 0.008349} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.856176] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0905707d-7c49-4a3c-9045-a5e9110bf052 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.860711] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Waiting for the task: (returnval){ [ 833.860711] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523274b2-5727-76ef-d543-5aa7d7932476" [ 833.860711] env[62952]: _type = "Task" [ 833.860711] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.867879] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523274b2-5727-76ef-d543-5aa7d7932476, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.874225] env[62952]: DEBUG nova.network.neutron [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 834.093303] env[62952]: DEBUG nova.network.neutron [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Updating instance_info_cache with network_info: [{"id": "363ca83f-c6ac-4446-996c-edc459eb29bf", "address": "fa:16:3e:dd:39:7e", "network": {"id": "3c430a3e-4ea2-4c7b-8e05-733f4a306f04", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-441856160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8960909bb80f4395bb87b6de3c60ab2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap363ca83f-c6", "ovs_interfaceid": "363ca83f-c6ac-4446-996c-edc459eb29bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.121200] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f2b5e5f-8117-48ac-833c-7479d09fca65 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.133235] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9d84036-043a-4c16-a70d-3cea3e31be41 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.166009] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e890ac0-8682-431c-97a3-973f78244a85 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.173529] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-508b0b57-59e8-483d-8bf0-b741ef5a5233 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.190709] env[62952]: DEBUG nova.compute.provider_tree [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.192635] env[62952]: DEBUG nova.network.neutron [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Successfully created port: b4e55382-c684-4ae1-afc7-0831ef60ac7e {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 834.202891] env[62952]: INFO nova.scheduler.client.report [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Deleted allocations for instance 1ed9de84-6178-4173-84ad-dc115be7a1fa [ 834.301703] env[62952]: DEBUG nova.compute.manager [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 834.371861] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523274b2-5727-76ef-d543-5aa7d7932476, 'name': SearchDatastore_Task, 'duration_secs': 0.008228} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.372138] env[62952]: DEBUG oslo_concurrency.lockutils [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.372394] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 8ac98d96-139a-4a03-94f3-178fafa88503/8ac98d96-139a-4a03-94f3-178fafa88503.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 834.372652] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f7091f3e-866a-48e4-b9be-8d8bc502945b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.379926] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Waiting for the task: (returnval){ [ 834.379926] env[62952]: value = "task-1263133" [ 834.379926] env[62952]: _type = "Task" [ 834.379926] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.387468] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': task-1263133, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.596384] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Releasing lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.596747] env[62952]: DEBUG nova.compute.manager [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Instance network_info: |[{"id": "363ca83f-c6ac-4446-996c-edc459eb29bf", "address": "fa:16:3e:dd:39:7e", "network": {"id": "3c430a3e-4ea2-4c7b-8e05-733f4a306f04", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-441856160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8960909bb80f4395bb87b6de3c60ab2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap363ca83f-c6", "ovs_interfaceid": "363ca83f-c6ac-4446-996c-edc459eb29bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 834.597697] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:39:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '03ac2c9c-6ad2-4a85-bfab-c7e336df859a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '363ca83f-c6ac-4446-996c-edc459eb29bf', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 834.606462] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Creating folder: Project (8960909bb80f4395bb87b6de3c60ab2c). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 834.606798] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9890fb4e-5b14-4318-9235-66450041a9a6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.617911] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Created folder: Project (8960909bb80f4395bb87b6de3c60ab2c) in parent group-v271811. [ 834.618174] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Creating folder: Instances. Parent ref: group-v271834. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 834.618445] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c60c1943-6db7-40de-a205-161d9117d8fa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.629490] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Created folder: Instances in parent group-v271834. [ 834.629743] env[62952]: DEBUG oslo.service.loopingcall [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.629943] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 834.630177] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0d91114d-acf4-41fb-abc1-66d9bd8c959b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.650747] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 834.650747] env[62952]: value = "task-1263136" [ 834.650747] env[62952]: _type = "Task" [ 834.650747] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.660510] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263136, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.695738] env[62952]: DEBUG nova.scheduler.client.report [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 834.711738] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbe91204-a9ab-4f29-b265-1afdf7f6ba6d tempest-ServerTagsTestJSON-1274640034 tempest-ServerTagsTestJSON-1274640034-project-member] Lock "1ed9de84-6178-4173-84ad-dc115be7a1fa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.509s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.890115] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': task-1263133, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.443157} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.890985] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 8ac98d96-139a-4a03-94f3-178fafa88503/8ac98d96-139a-4a03-94f3-178fafa88503.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 834.890985] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.891197] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5bfa1968-cad8-44db-af37-4a1c24bcb549 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.897851] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Waiting for the task: (returnval){ [ 834.897851] env[62952]: value = "task-1263137" [ 834.897851] env[62952]: _type = "Task" [ 834.897851] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.908663] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': task-1263137, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.160242] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263136, 'name': CreateVM_Task, 'duration_secs': 0.38115} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.160484] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 835.161172] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.161343] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.161746] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 835.161994] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5f0c120-78f0-4a64-bdc0-9748db288047 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.166800] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 835.166800] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52aabfbb-06af-afdc-7cde-6866ddf4df5d" [ 835.166800] env[62952]: _type = "Task" [ 835.166800] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.174948] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52aabfbb-06af-afdc-7cde-6866ddf4df5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.201756] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.913s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.202339] env[62952]: ERROR nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 559b10ed-15b2-417c-bf21-939af6239155, please check neutron logs for more information. [ 835.202339] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Traceback (most recent call last): [ 835.202339] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 835.202339] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] self.driver.spawn(context, instance, image_meta, [ 835.202339] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 835.202339] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 835.202339] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 835.202339] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] vm_ref = self.build_virtual_machine(instance, [ 835.202339] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 835.202339] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] vif_infos = vmwarevif.get_vif_info(self._session, [ 835.202339] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 835.202740] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] for vif in network_info: [ 835.202740] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 835.202740] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] return self._sync_wrapper(fn, *args, **kwargs) [ 835.202740] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 835.202740] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] self.wait() [ 835.202740] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 835.202740] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] self[:] = self._gt.wait() [ 835.202740] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 835.202740] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] return self._exit_event.wait() [ 835.202740] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 835.202740] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] current.throw(*self._exc) [ 835.202740] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 835.202740] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] result = function(*args, **kwargs) [ 835.203169] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 835.203169] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] return func(*args, **kwargs) [ 835.203169] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 835.203169] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] raise e [ 835.203169] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 835.203169] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] nwinfo = self.network_api.allocate_for_instance( [ 835.203169] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 835.203169] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] created_port_ids = self._update_ports_for_instance( [ 835.203169] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 835.203169] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] with excutils.save_and_reraise_exception(): [ 835.203169] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 835.203169] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] self.force_reraise() [ 835.203169] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 835.203588] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] raise self.value [ 835.203588] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 835.203588] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] updated_port = self._update_port( [ 835.203588] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 835.203588] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] _ensure_no_port_binding_failure(port) [ 835.203588] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 835.203588] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] raise exception.PortBindingFailed(port_id=port['id']) [ 835.203588] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] nova.exception.PortBindingFailed: Binding failed for port 559b10ed-15b2-417c-bf21-939af6239155, please check neutron logs for more information. [ 835.203588] env[62952]: ERROR nova.compute.manager [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] [ 835.203588] env[62952]: DEBUG nova.compute.utils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Binding failed for port 559b10ed-15b2-417c-bf21-939af6239155, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 835.204397] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.830s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.205878] env[62952]: INFO nova.compute.claims [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 835.208500] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Build of instance 1d6447ab-dc07-459f-9124-d881e0da7cf1 was re-scheduled: Binding failed for port 559b10ed-15b2-417c-bf21-939af6239155, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 835.209999] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 835.210237] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "refresh_cache-1d6447ab-dc07-459f-9124-d881e0da7cf1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.210435] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquired lock "refresh_cache-1d6447ab-dc07-459f-9124-d881e0da7cf1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.210608] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 835.217234] env[62952]: DEBUG nova.compute.manager [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 835.310689] env[62952]: DEBUG nova.compute.manager [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 835.335843] env[62952]: DEBUG nova.virt.hardware [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 835.336102] env[62952]: DEBUG nova.virt.hardware [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 835.336261] env[62952]: DEBUG nova.virt.hardware [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.336434] env[62952]: DEBUG nova.virt.hardware [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 835.336610] env[62952]: DEBUG nova.virt.hardware [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.336759] env[62952]: DEBUG nova.virt.hardware [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 835.336960] env[62952]: DEBUG nova.virt.hardware [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 835.337142] env[62952]: DEBUG nova.virt.hardware [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 835.337305] env[62952]: DEBUG nova.virt.hardware [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 835.337466] env[62952]: DEBUG nova.virt.hardware [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 835.337630] env[62952]: DEBUG nova.virt.hardware [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.338514] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-206fef14-b12f-4614-88a7-ca343bde57e8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.346481] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-486c5e0e-9de9-450b-bf84-4192dc7ee41d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.407750] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': task-1263137, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061249} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.407927] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 835.408763] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f9aa54-9529-4f3e-942b-a02c8f3fa2cb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.433033] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] 8ac98d96-139a-4a03-94f3-178fafa88503/8ac98d96-139a-4a03-94f3-178fafa88503.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.433311] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9592ff4a-c097-4679-91f9-ca9cff19ddaf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.452061] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Waiting for the task: (returnval){ [ 835.452061] env[62952]: value = "task-1263138" [ 835.452061] env[62952]: _type = "Task" [ 835.452061] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.461961] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': task-1263138, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.646485] env[62952]: DEBUG nova.compute.manager [req-eb640d51-0610-4202-8aa3-64d3fe87bcbf req-e0517aad-ba32-47cc-90bc-eb4222b49bd1 service nova] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Received event network-vif-plugged-b4e55382-c684-4ae1-afc7-0831ef60ac7e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 835.646752] env[62952]: DEBUG oslo_concurrency.lockutils [req-eb640d51-0610-4202-8aa3-64d3fe87bcbf req-e0517aad-ba32-47cc-90bc-eb4222b49bd1 service nova] Acquiring lock "5f1eca6f-4cc4-483e-b72a-b479378277fe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.646959] env[62952]: DEBUG oslo_concurrency.lockutils [req-eb640d51-0610-4202-8aa3-64d3fe87bcbf req-e0517aad-ba32-47cc-90bc-eb4222b49bd1 service nova] Lock "5f1eca6f-4cc4-483e-b72a-b479378277fe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.647199] env[62952]: DEBUG oslo_concurrency.lockutils [req-eb640d51-0610-4202-8aa3-64d3fe87bcbf req-e0517aad-ba32-47cc-90bc-eb4222b49bd1 service nova] Lock "5f1eca6f-4cc4-483e-b72a-b479378277fe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.647402] env[62952]: DEBUG nova.compute.manager [req-eb640d51-0610-4202-8aa3-64d3fe87bcbf req-e0517aad-ba32-47cc-90bc-eb4222b49bd1 service nova] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] No waiting events found dispatching network-vif-plugged-b4e55382-c684-4ae1-afc7-0831ef60ac7e {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 835.647599] env[62952]: WARNING nova.compute.manager [req-eb640d51-0610-4202-8aa3-64d3fe87bcbf req-e0517aad-ba32-47cc-90bc-eb4222b49bd1 service nova] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Received unexpected event network-vif-plugged-b4e55382-c684-4ae1-afc7-0831ef60ac7e for instance with vm_state building and task_state spawning. [ 835.676896] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52aabfbb-06af-afdc-7cde-6866ddf4df5d, 'name': SearchDatastore_Task, 'duration_secs': 0.01177} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.677373] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.677643] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 835.677904] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.678090] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.678380] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.678673] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c518692-2b31-459a-a524-87254a4a1fd2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.686351] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.686522] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 835.687362] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-524d9730-89c1-4a0f-aa45-10dbcce0e1d9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.692375] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 835.692375] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521f5c9f-5624-1b6d-98ea-3ec16c858935" [ 835.692375] env[62952]: _type = "Task" [ 835.692375] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.700057] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521f5c9f-5624-1b6d-98ea-3ec16c858935, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.720585] env[62952]: DEBUG nova.network.neutron [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Successfully updated port: b4e55382-c684-4ae1-afc7-0831ef60ac7e {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 835.743251] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 835.748429] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.835292] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.965061] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': task-1263138, 'name': ReconfigVM_Task, 'duration_secs': 0.289527} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.965061] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Reconfigured VM instance instance-00000030 to attach disk [datastore2] 8ac98d96-139a-4a03-94f3-178fafa88503/8ac98d96-139a-4a03-94f3-178fafa88503.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.965673] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a0bda49e-205e-42c9-8d16-ab6b0b3e95d9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.972116] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Waiting for the task: (returnval){ [ 835.972116] env[62952]: value = "task-1263139" [ 835.972116] env[62952]: _type = "Task" [ 835.972116] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.980118] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': task-1263139, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.202997] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521f5c9f-5624-1b6d-98ea-3ec16c858935, 'name': SearchDatastore_Task, 'duration_secs': 0.009867} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.203782] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30c72031-7c98-40a2-a7c1-74cbd5022c78 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.208722] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 836.208722] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52afee1b-57a0-ba4e-a798-8b3a6b7191b7" [ 836.208722] env[62952]: _type = "Task" [ 836.208722] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.216407] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52afee1b-57a0-ba4e-a798-8b3a6b7191b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.227967] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "refresh_cache-5f1eca6f-4cc4-483e-b72a-b479378277fe" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.228156] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquired lock "refresh_cache-5f1eca6f-4cc4-483e-b72a-b479378277fe" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.228263] env[62952]: DEBUG nova.network.neutron [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 836.337584] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Releasing lock "refresh_cache-1d6447ab-dc07-459f-9124-d881e0da7cf1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.337929] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 836.338735] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 836.338735] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 836.354144] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 836.483484] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': task-1263139, 'name': Rename_Task, 'duration_secs': 0.175932} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.483767] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 836.484016] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0bfa2ea6-d268-4900-ab29-8561e6db7b9b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.491779] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Waiting for the task: (returnval){ [ 836.491779] env[62952]: value = "task-1263140" [ 836.491779] env[62952]: _type = "Task" [ 836.491779] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.499867] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': task-1263140, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.524777] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56bebb7-9ea3-479f-9d64-7732ad147704 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.532270] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df675871-29d3-4aa2-baff-e2f575d1ac87 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.562519] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc1d5523-4356-4105-b4b3-b7b052606ce8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.570665] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a00189e-d0ab-407c-b787-3800a03385a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.585807] env[62952]: DEBUG nova.compute.provider_tree [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 836.718965] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52afee1b-57a0-ba4e-a798-8b3a6b7191b7, 'name': SearchDatastore_Task, 'duration_secs': 0.008336} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.719251] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.719503] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] f4f8bee4-bf92-4321-afdc-16f7fb17dbae/f4f8bee4-bf92-4321-afdc-16f7fb17dbae.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 836.719756] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-24f402e6-cc13-4576-8e14-65b059643144 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.726327] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 836.726327] env[62952]: value = "task-1263141" [ 836.726327] env[62952]: _type = "Task" [ 836.726327] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.737795] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263141, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.771594] env[62952]: DEBUG nova.network.neutron [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 836.858244] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.966731] env[62952]: DEBUG nova.network.neutron [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Updating instance_info_cache with network_info: [{"id": "b4e55382-c684-4ae1-afc7-0831ef60ac7e", "address": "fa:16:3e:d9:71:3a", "network": {"id": "2c5146eb-dcff-470d-984f-b96a3c91bb3e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1950178234-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1487c06883e444a589d9b688155dc00e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4e55382-c6", "ovs_interfaceid": "b4e55382-c684-4ae1-afc7-0831ef60ac7e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.004849] env[62952]: DEBUG oslo_vmware.api [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': task-1263140, 'name': PowerOnVM_Task, 'duration_secs': 0.462779} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.005340] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 837.005722] env[62952]: INFO nova.compute.manager [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Took 8.41 seconds to spawn the instance on the hypervisor. [ 837.006069] env[62952]: DEBUG nova.compute.manager [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.007025] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce95338a-259e-4069-8223-c919a088bbbe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.108840] env[62952]: ERROR nova.scheduler.client.report [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [req-3c70b503-f1af-40ce-9373-fa888f3b44b8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 615127b5-dc82-4491-9c8f-4e6d0caa0690. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3c70b503-f1af-40ce-9373-fa888f3b44b8"}]} [ 837.126720] env[62952]: DEBUG nova.scheduler.client.report [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Refreshing inventories for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 837.141692] env[62952]: DEBUG nova.scheduler.client.report [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Updating ProviderTree inventory for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 837.141942] env[62952]: DEBUG nova.compute.provider_tree [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 837.154063] env[62952]: DEBUG nova.scheduler.client.report [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Refreshing aggregate associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 837.174161] env[62952]: DEBUG nova.scheduler.client.report [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Refreshing trait associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 837.236833] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263141, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458324} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.239366] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] f4f8bee4-bf92-4321-afdc-16f7fb17dbae/f4f8bee4-bf92-4321-afdc-16f7fb17dbae.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 837.239584] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.240074] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c696618d-7cb2-47cb-bb91-bc9abcb57bb8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.246908] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 837.246908] env[62952]: value = "task-1263142" [ 837.246908] env[62952]: _type = "Task" [ 837.246908] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.256328] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263142, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.364092] env[62952]: INFO nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 1d6447ab-dc07-459f-9124-d881e0da7cf1] Took 1.03 seconds to deallocate network for instance. [ 837.469466] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Releasing lock "refresh_cache-5f1eca6f-4cc4-483e-b72a-b479378277fe" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.469523] env[62952]: DEBUG nova.compute.manager [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Instance network_info: |[{"id": "b4e55382-c684-4ae1-afc7-0831ef60ac7e", "address": "fa:16:3e:d9:71:3a", "network": {"id": "2c5146eb-dcff-470d-984f-b96a3c91bb3e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1950178234-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1487c06883e444a589d9b688155dc00e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4e55382-c6", "ovs_interfaceid": "b4e55382-c684-4ae1-afc7-0831ef60ac7e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 837.470806] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:71:3a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e445fb59-822c-4d7d-943b-c8e3bbaca62e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b4e55382-c684-4ae1-afc7-0831ef60ac7e', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.478265] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Creating folder: Project (1487c06883e444a589d9b688155dc00e). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 837.479353] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cdcaf33-3e98-4280-afe6-657d9c35ca91 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.482311] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a454614b-3dd5-4698-a3d0-9ed6d8b33fdb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.488366] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b4ae0c8-310b-489f-8fdc-321787081f6f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.492450] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Created folder: Project (1487c06883e444a589d9b688155dc00e) in parent group-v271811. [ 837.492622] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Creating folder: Instances. Parent ref: group-v271837. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 837.493111] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4bd275ac-d311-48f9-ba59-77432b470830 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.526117] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b00300a-2a8b-40e4-a272-b337827fb2cd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.527584] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Created folder: Instances in parent group-v271837. [ 837.527804] env[62952]: DEBUG oslo.service.loopingcall [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.529482] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 837.530026] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ced7be97-d2ec-4fc3-aa64-d5c11b8eb3eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.544035] env[62952]: INFO nova.compute.manager [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Took 33.72 seconds to build instance. [ 837.548292] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77424ab-3a4d-4195-8297-3b9c5a8f4155 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.552846] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.552846] env[62952]: value = "task-1263145" [ 837.552846] env[62952]: _type = "Task" [ 837.552846] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.563336] env[62952]: DEBUG nova.compute.provider_tree [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 837.569643] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263145, 'name': CreateVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.683970] env[62952]: DEBUG nova.compute.manager [req-efee2fca-7bd0-4e3b-8d62-269818e9d182 req-bacf7f18-22fc-4e54-ae40-4a9341a4d48d service nova] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Received event network-changed-b4e55382-c684-4ae1-afc7-0831ef60ac7e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 837.684228] env[62952]: DEBUG nova.compute.manager [req-efee2fca-7bd0-4e3b-8d62-269818e9d182 req-bacf7f18-22fc-4e54-ae40-4a9341a4d48d service nova] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Refreshing instance network info cache due to event network-changed-b4e55382-c684-4ae1-afc7-0831ef60ac7e. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 837.685060] env[62952]: DEBUG oslo_concurrency.lockutils [req-efee2fca-7bd0-4e3b-8d62-269818e9d182 req-bacf7f18-22fc-4e54-ae40-4a9341a4d48d service nova] Acquiring lock "refresh_cache-5f1eca6f-4cc4-483e-b72a-b479378277fe" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.685060] env[62952]: DEBUG oslo_concurrency.lockutils [req-efee2fca-7bd0-4e3b-8d62-269818e9d182 req-bacf7f18-22fc-4e54-ae40-4a9341a4d48d service nova] Acquired lock "refresh_cache-5f1eca6f-4cc4-483e-b72a-b479378277fe" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.685060] env[62952]: DEBUG nova.network.neutron [req-efee2fca-7bd0-4e3b-8d62-269818e9d182 req-bacf7f18-22fc-4e54-ae40-4a9341a4d48d service nova] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Refreshing network info cache for port b4e55382-c684-4ae1-afc7-0831ef60ac7e {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 837.757652] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263142, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077873} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.757925] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 837.758843] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa7da39-ae2e-4ede-936a-334576b4d179 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.781908] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] f4f8bee4-bf92-4321-afdc-16f7fb17dbae/f4f8bee4-bf92-4321-afdc-16f7fb17dbae.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.782410] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0d16c61-4edd-47db-9605-90cd15b0ab1d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.802192] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 837.802192] env[62952]: value = "task-1263146" [ 837.802192] env[62952]: _type = "Task" [ 837.802192] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.810335] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263146, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.049289] env[62952]: DEBUG oslo_concurrency.lockutils [None req-945b7aa2-3375-4c74-bab6-0fd54e3c7373 tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Lock "8ac98d96-139a-4a03-94f3-178fafa88503" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.706s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.062722] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263145, 'name': CreateVM_Task, 'duration_secs': 0.346377} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.063494] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 838.064226] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.064402] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.064710] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 838.065197] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f792fb47-a1e1-44a3-a89c-18d0c4957dc0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.072994] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 838.072994] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52052576-6180-6d2b-7636-cf260a14d795" [ 838.072994] env[62952]: _type = "Task" [ 838.072994] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.081689] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52052576-6180-6d2b-7636-cf260a14d795, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.086088] env[62952]: ERROR nova.scheduler.client.report [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [req-3b812308-900c-47ee-9dd6-a6fed6918ca2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 615127b5-dc82-4491-9c8f-4e6d0caa0690. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3b812308-900c-47ee-9dd6-a6fed6918ca2"}]} [ 838.101497] env[62952]: DEBUG nova.scheduler.client.report [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Refreshing inventories for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 838.116667] env[62952]: DEBUG nova.scheduler.client.report [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Updating ProviderTree inventory for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 838.116877] env[62952]: DEBUG nova.compute.provider_tree [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 838.127544] env[62952]: DEBUG nova.scheduler.client.report [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Refreshing aggregate associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 838.143996] env[62952]: DEBUG nova.scheduler.client.report [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Refreshing trait associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 838.284413] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "86ec7b3e-34d6-45e8-87b2-c8550f035ed0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.284413] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "86ec7b3e-34d6-45e8-87b2-c8550f035ed0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.314478] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263146, 'name': ReconfigVM_Task, 'duration_secs': 0.262256} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.317020] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Reconfigured VM instance instance-00000031 to attach disk [datastore2] f4f8bee4-bf92-4321-afdc-16f7fb17dbae/f4f8bee4-bf92-4321-afdc-16f7fb17dbae.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 838.318055] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d6bf5d96-f21b-4fbc-83c2-96f641922b35 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.324483] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 838.324483] env[62952]: value = "task-1263147" [ 838.324483] env[62952]: _type = "Task" [ 838.324483] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.335082] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263147, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.397259] env[62952]: INFO nova.scheduler.client.report [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Deleted allocations for instance 1d6447ab-dc07-459f-9124-d881e0da7cf1 [ 838.444068] env[62952]: DEBUG nova.network.neutron [req-efee2fca-7bd0-4e3b-8d62-269818e9d182 req-bacf7f18-22fc-4e54-ae40-4a9341a4d48d service nova] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Updated VIF entry in instance network info cache for port b4e55382-c684-4ae1-afc7-0831ef60ac7e. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 838.444068] env[62952]: DEBUG nova.network.neutron [req-efee2fca-7bd0-4e3b-8d62-269818e9d182 req-bacf7f18-22fc-4e54-ae40-4a9341a4d48d service nova] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Updating instance_info_cache with network_info: [{"id": "b4e55382-c684-4ae1-afc7-0831ef60ac7e", "address": "fa:16:3e:d9:71:3a", "network": {"id": "2c5146eb-dcff-470d-984f-b96a3c91bb3e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1950178234-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1487c06883e444a589d9b688155dc00e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4e55382-c6", "ovs_interfaceid": "b4e55382-c684-4ae1-afc7-0831ef60ac7e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.469223] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Acquiring lock "8ac98d96-139a-4a03-94f3-178fafa88503" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.469336] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Lock "8ac98d96-139a-4a03-94f3-178fafa88503" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.470027] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Acquiring lock "8ac98d96-139a-4a03-94f3-178fafa88503-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.470027] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Lock "8ac98d96-139a-4a03-94f3-178fafa88503-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.470027] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Lock "8ac98d96-139a-4a03-94f3-178fafa88503-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.471808] env[62952]: INFO nova.compute.manager [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Terminating instance [ 838.474897] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc637c20-0946-4ffa-803d-018e4370de59 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.478096] env[62952]: DEBUG nova.compute.manager [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 838.478285] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 838.479017] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e760dce9-320e-4d85-a822-919b98178e8b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.487032] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1eff680-e11f-4852-9dba-295b667000f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.490038] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 838.490476] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-69bb0524-3d1c-40b3-8ee2-6a1847c1cb2b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.521881] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d59df7-7357-4b5d-bc24-daa5b5c31746 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.526773] env[62952]: DEBUG oslo_vmware.api [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Waiting for the task: (returnval){ [ 838.526773] env[62952]: value = "task-1263148" [ 838.526773] env[62952]: _type = "Task" [ 838.526773] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.527217] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "47b57348-40d1-4735-9f5a-a7288f8f12ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.527419] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "47b57348-40d1-4735-9f5a-a7288f8f12ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.534020] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c749f1cc-68b0-448c-8029-7eaad0eb1625 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.540372] env[62952]: DEBUG oslo_vmware.api [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': task-1263148, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.549914] env[62952]: DEBUG nova.compute.provider_tree [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 838.551313] env[62952]: DEBUG nova.compute.manager [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 838.583861] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52052576-6180-6d2b-7636-cf260a14d795, 'name': SearchDatastore_Task, 'duration_secs': 0.008242} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.584181] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.584408] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.584624] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.584763] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.585631] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.585631] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d135ae3f-79fc-4189-9b7d-8098a16a6fa3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.593124] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.593306] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 838.594037] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbbe29d5-3bea-49ba-bb57-68f789b62de2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.599759] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 838.599759] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5282c01a-316b-18f8-4877-dcd265e17e7d" [ 838.599759] env[62952]: _type = "Task" [ 838.599759] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.607462] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5282c01a-316b-18f8-4877-dcd265e17e7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.833638] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263147, 'name': Rename_Task, 'duration_secs': 0.132032} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.833915] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 838.834170] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-68d9bb2a-5559-421e-96a3-eb9e80ca9ca2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.840012] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 838.840012] env[62952]: value = "task-1263149" [ 838.840012] env[62952]: _type = "Task" [ 838.840012] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.846895] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263149, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.909067] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "1d6447ab-dc07-459f-9124-d881e0da7cf1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.279s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.946460] env[62952]: DEBUG oslo_concurrency.lockutils [req-efee2fca-7bd0-4e3b-8d62-269818e9d182 req-bacf7f18-22fc-4e54-ae40-4a9341a4d48d service nova] Releasing lock "refresh_cache-5f1eca6f-4cc4-483e-b72a-b479378277fe" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.037246] env[62952]: DEBUG oslo_vmware.api [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': task-1263148, 'name': PowerOffVM_Task, 'duration_secs': 0.270254} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.037512] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 839.037679] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 839.037922] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-07190478-8db1-4f5a-8029-d13ff03b5dce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.074683] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.088084] env[62952]: DEBUG nova.scheduler.client.report [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Updated inventory for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with generation 80 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 839.088376] env[62952]: DEBUG nova.compute.provider_tree [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Updating resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 generation from 80 to 81 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 839.088709] env[62952]: DEBUG nova.compute.provider_tree [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 839.107989] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 839.107989] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 839.107989] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Deleting the datastore file [datastore2] 8ac98d96-139a-4a03-94f3-178fafa88503 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 839.108772] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0ae0f965-7d00-465b-91bf-17f98b5f9537 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.115261] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5282c01a-316b-18f8-4877-dcd265e17e7d, 'name': SearchDatastore_Task, 'duration_secs': 0.008896} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.116644] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12fefdc6-a47d-4918-bd79-bb002d251a1c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.120211] env[62952]: DEBUG oslo_vmware.api [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Waiting for the task: (returnval){ [ 839.120211] env[62952]: value = "task-1263151" [ 839.120211] env[62952]: _type = "Task" [ 839.120211] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.124613] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 839.124613] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e97f80-8738-3de8-4225-fb9435163b9d" [ 839.124613] env[62952]: _type = "Task" [ 839.124613] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.130318] env[62952]: DEBUG oslo_vmware.api [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': task-1263151, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.134823] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e97f80-8738-3de8-4225-fb9435163b9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.349812] env[62952]: DEBUG oslo_vmware.api [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263149, 'name': PowerOnVM_Task, 'duration_secs': 0.432722} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.350061] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 839.350922] env[62952]: INFO nova.compute.manager [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Took 8.33 seconds to spawn the instance on the hypervisor. [ 839.350922] env[62952]: DEBUG nova.compute.manager [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.351284] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee661ea1-4aba-4519-b288-097800e5fdf8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.412224] env[62952]: DEBUG nova.compute.manager [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 839.595050] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.390s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.595050] env[62952]: DEBUG nova.compute.manager [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 839.597511] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.387s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.599134] env[62952]: INFO nova.compute.claims [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 839.632436] env[62952]: DEBUG oslo_vmware.api [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Task: {'id': task-1263151, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151695} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.633140] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 839.633380] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 839.633572] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 839.633744] env[62952]: INFO nova.compute.manager [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Took 1.16 seconds to destroy the instance on the hypervisor. [ 839.634056] env[62952]: DEBUG oslo.service.loopingcall [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 839.637495] env[62952]: DEBUG nova.compute.manager [-] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 839.637611] env[62952]: DEBUG nova.network.neutron [-] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 839.639246] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e97f80-8738-3de8-4225-fb9435163b9d, 'name': SearchDatastore_Task, 'duration_secs': 0.016981} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.639488] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.639728] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 5f1eca6f-4cc4-483e-b72a-b479378277fe/5f1eca6f-4cc4-483e-b72a-b479378277fe.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 839.641038] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-02aac7b7-854d-4d5a-9cb2-449322e1a942 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.646922] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 839.646922] env[62952]: value = "task-1263152" [ 839.646922] env[62952]: _type = "Task" [ 839.646922] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.655117] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263152, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.869474] env[62952]: INFO nova.compute.manager [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Took 34.17 seconds to build instance. [ 839.934031] env[62952]: DEBUG nova.compute.manager [req-416679dc-7320-4691-9e8f-22159715ad54 req-49059c5f-f0f1-446b-98f0-de1f685e0dfd service nova] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Received event network-vif-deleted-6af48a64-0373-41eb-a15f-976266b85305 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 839.934085] env[62952]: INFO nova.compute.manager [req-416679dc-7320-4691-9e8f-22159715ad54 req-49059c5f-f0f1-446b-98f0-de1f685e0dfd service nova] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Neutron deleted interface 6af48a64-0373-41eb-a15f-976266b85305; detaching it from the instance and deleting it from the info cache [ 839.934614] env[62952]: DEBUG nova.network.neutron [req-416679dc-7320-4691-9e8f-22159715ad54 req-49059c5f-f0f1-446b-98f0-de1f685e0dfd service nova] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.938911] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.103705] env[62952]: DEBUG nova.compute.utils [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 840.107494] env[62952]: DEBUG nova.compute.manager [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 840.107652] env[62952]: DEBUG nova.network.neutron [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 840.150045] env[62952]: DEBUG nova.policy [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '42a603de55734dff95126c1c9bcc8699', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd93803f89ca64e66b8271e7b0bb48983', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 840.161929] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263152, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463705} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.162292] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 5f1eca6f-4cc4-483e-b72a-b479378277fe/5f1eca6f-4cc4-483e-b72a-b479378277fe.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 840.162517] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.162782] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dd41beda-af05-4c26-8b20-4703055d5dab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.168939] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 840.168939] env[62952]: value = "task-1263153" [ 840.168939] env[62952]: _type = "Task" [ 840.168939] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.176148] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263153, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.372635] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f4f7741c-3770-4ea7-93b9-1bb38a40f5b4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.856s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.409609] env[62952]: DEBUG nova.network.neutron [-] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.426751] env[62952]: DEBUG nova.network.neutron [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Successfully created port: 3a705583-6448-4020-9894-5f2056e2e4e7 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 840.436633] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-59bb3cea-3ea4-450f-96fb-a21c3489bb40 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.445416] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26ec85d-b17a-463c-b379-927ae0ea30bc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.471069] env[62952]: DEBUG nova.compute.manager [req-416679dc-7320-4691-9e8f-22159715ad54 req-49059c5f-f0f1-446b-98f0-de1f685e0dfd service nova] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Detach interface failed, port_id=6af48a64-0373-41eb-a15f-976266b85305, reason: Instance 8ac98d96-139a-4a03-94f3-178fafa88503 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 840.608500] env[62952]: DEBUG nova.compute.manager [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 840.680408] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263153, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058189} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.680740] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 840.681722] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b79d7b64-cc85-424d-b1c5-aa8161a6a242 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.705228] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 5f1eca6f-4cc4-483e-b72a-b479378277fe/5f1eca6f-4cc4-483e-b72a-b479378277fe.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.707864] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c838d36-730c-4bce-9671-9c8f0cb26607 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.727518] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 840.727518] env[62952]: value = "task-1263154" [ 840.727518] env[62952]: _type = "Task" [ 840.727518] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.741086] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263154, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.876028] env[62952]: DEBUG nova.compute.manager [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 840.912035] env[62952]: INFO nova.compute.manager [-] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Took 1.27 seconds to deallocate network for instance. [ 840.959156] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b2be0d-0273-47b1-8d69-d4fc0211b4f8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.967579] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548f2f55-8785-4727-8b27-c18c005fc854 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.000115] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061221b2-479c-4772-a816-52ef0437e35f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.008234] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac19672c-2f7f-4db4-9423-6afa30541b37 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.022053] env[62952]: DEBUG nova.compute.provider_tree [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.237626] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263154, 'name': ReconfigVM_Task, 'duration_secs': 0.302095} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.237904] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 5f1eca6f-4cc4-483e-b72a-b479378277fe/5f1eca6f-4cc4-483e-b72a-b479378277fe.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.238534] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ec1d9969-56d5-4c7f-8592-b30c27c0be25 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.245650] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 841.245650] env[62952]: value = "task-1263155" [ 841.245650] env[62952]: _type = "Task" [ 841.245650] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.253129] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263155, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.406092] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.419454] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.525764] env[62952]: DEBUG nova.scheduler.client.report [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 841.585466] env[62952]: DEBUG nova.compute.manager [req-6c2047eb-8139-4d6b-98b9-e8175f6f001a req-d76a6af4-eeaa-486b-b2b6-727c498573c7 service nova] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Received event network-changed-363ca83f-c6ac-4446-996c-edc459eb29bf {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 841.585560] env[62952]: DEBUG nova.compute.manager [req-6c2047eb-8139-4d6b-98b9-e8175f6f001a req-d76a6af4-eeaa-486b-b2b6-727c498573c7 service nova] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Refreshing instance network info cache due to event network-changed-363ca83f-c6ac-4446-996c-edc459eb29bf. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 841.585870] env[62952]: DEBUG oslo_concurrency.lockutils [req-6c2047eb-8139-4d6b-98b9-e8175f6f001a req-d76a6af4-eeaa-486b-b2b6-727c498573c7 service nova] Acquiring lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.585990] env[62952]: DEBUG oslo_concurrency.lockutils [req-6c2047eb-8139-4d6b-98b9-e8175f6f001a req-d76a6af4-eeaa-486b-b2b6-727c498573c7 service nova] Acquired lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.586091] env[62952]: DEBUG nova.network.neutron [req-6c2047eb-8139-4d6b-98b9-e8175f6f001a req-d76a6af4-eeaa-486b-b2b6-727c498573c7 service nova] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Refreshing network info cache for port 363ca83f-c6ac-4446-996c-edc459eb29bf {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 841.621431] env[62952]: DEBUG nova.compute.manager [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 841.655355] env[62952]: DEBUG nova.virt.hardware [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 841.655649] env[62952]: DEBUG nova.virt.hardware [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 841.655839] env[62952]: DEBUG nova.virt.hardware [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 841.656152] env[62952]: DEBUG nova.virt.hardware [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 841.656351] env[62952]: DEBUG nova.virt.hardware [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 841.656536] env[62952]: DEBUG nova.virt.hardware [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 841.657230] env[62952]: DEBUG nova.virt.hardware [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 841.657230] env[62952]: DEBUG nova.virt.hardware [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 841.657346] env[62952]: DEBUG nova.virt.hardware [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 841.657445] env[62952]: DEBUG nova.virt.hardware [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 841.657626] env[62952]: DEBUG nova.virt.hardware [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 841.658971] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d614734-e8d9-4fe6-a8d8-8f51b527b2ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.669293] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b317f2-68b6-43d6-a7d9-a2ed867eef75 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.757397] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263155, 'name': Rename_Task, 'duration_secs': 0.129627} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.757705] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 841.757971] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-976ea48b-867a-48e8-a450-0f565750f422 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.765874] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 841.765874] env[62952]: value = "task-1263156" [ 841.765874] env[62952]: _type = "Task" [ 841.765874] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.775939] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263156, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.968534] env[62952]: DEBUG nova.compute.manager [req-e8917f1d-f301-4897-b7ea-9ed25b7ac772 req-b7e663ac-894a-43df-82b6-d9dfb1c4a025 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Received event network-vif-plugged-3a705583-6448-4020-9894-5f2056e2e4e7 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 841.968767] env[62952]: DEBUG oslo_concurrency.lockutils [req-e8917f1d-f301-4897-b7ea-9ed25b7ac772 req-b7e663ac-894a-43df-82b6-d9dfb1c4a025 service nova] Acquiring lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.969013] env[62952]: DEBUG oslo_concurrency.lockutils [req-e8917f1d-f301-4897-b7ea-9ed25b7ac772 req-b7e663ac-894a-43df-82b6-d9dfb1c4a025 service nova] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.969139] env[62952]: DEBUG oslo_concurrency.lockutils [req-e8917f1d-f301-4897-b7ea-9ed25b7ac772 req-b7e663ac-894a-43df-82b6-d9dfb1c4a025 service nova] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.969328] env[62952]: DEBUG nova.compute.manager [req-e8917f1d-f301-4897-b7ea-9ed25b7ac772 req-b7e663ac-894a-43df-82b6-d9dfb1c4a025 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] No waiting events found dispatching network-vif-plugged-3a705583-6448-4020-9894-5f2056e2e4e7 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 841.969460] env[62952]: WARNING nova.compute.manager [req-e8917f1d-f301-4897-b7ea-9ed25b7ac772 req-b7e663ac-894a-43df-82b6-d9dfb1c4a025 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Received unexpected event network-vif-plugged-3a705583-6448-4020-9894-5f2056e2e4e7 for instance with vm_state building and task_state spawning. [ 841.971612] env[62952]: DEBUG nova.network.neutron [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Successfully updated port: 3a705583-6448-4020-9894-5f2056e2e4e7 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 842.033427] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.436s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.033965] env[62952]: DEBUG nova.compute.manager [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 842.036941] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 23.583s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.037144] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.037303] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 842.037592] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.935s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.041533] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b012e6b-7583-4656-8502-02b618c780f3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.050549] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ba6541-b867-46eb-9c9f-5c2396879deb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.066067] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed6beec7-bd89-4841-bd7a-9014677f07e3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.071954] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf280f0-0bf7-489b-99f0-ce6a43fbb3a7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.105404] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181527MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 842.105610] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.276018] env[62952]: DEBUG oslo_vmware.api [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263156, 'name': PowerOnVM_Task, 'duration_secs': 0.439129} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.276355] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 842.276559] env[62952]: INFO nova.compute.manager [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Took 6.97 seconds to spawn the instance on the hypervisor. [ 842.276735] env[62952]: DEBUG nova.compute.manager [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 842.277583] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db69417b-68bc-42a3-8430-f4cb5dedeaea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.329477] env[62952]: DEBUG nova.network.neutron [req-6c2047eb-8139-4d6b-98b9-e8175f6f001a req-d76a6af4-eeaa-486b-b2b6-727c498573c7 service nova] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Updated VIF entry in instance network info cache for port 363ca83f-c6ac-4446-996c-edc459eb29bf. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 842.329858] env[62952]: DEBUG nova.network.neutron [req-6c2047eb-8139-4d6b-98b9-e8175f6f001a req-d76a6af4-eeaa-486b-b2b6-727c498573c7 service nova] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Updating instance_info_cache with network_info: [{"id": "363ca83f-c6ac-4446-996c-edc459eb29bf", "address": "fa:16:3e:dd:39:7e", "network": {"id": "3c430a3e-4ea2-4c7b-8e05-733f4a306f04", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-441856160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8960909bb80f4395bb87b6de3c60ab2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap363ca83f-c6", "ovs_interfaceid": "363ca83f-c6ac-4446-996c-edc459eb29bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.474236] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.474593] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.474593] env[62952]: DEBUG nova.network.neutron [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 842.542537] env[62952]: DEBUG nova.compute.utils [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 842.547034] env[62952]: DEBUG nova.compute.manager [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 842.547216] env[62952]: DEBUG nova.network.neutron [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 842.586123] env[62952]: DEBUG nova.policy [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4f4272707d14a959d63ad36f5f772a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e698fc6474024bbca26d6b6c06082f03', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 842.796841] env[62952]: INFO nova.compute.manager [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Took 31.70 seconds to build instance. [ 842.833609] env[62952]: DEBUG oslo_concurrency.lockutils [req-6c2047eb-8139-4d6b-98b9-e8175f6f001a req-d76a6af4-eeaa-486b-b2b6-727c498573c7 service nova] Releasing lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.864885] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01f979a-5f46-4822-bae9-1fbf91244c5f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.872974] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0703b36d-2a33-4c9d-aca9-bf305705b3a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.907394] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109b7bcd-33c9-4d5c-873b-fa4ebb2b5425 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.915337] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22a8a5d3-0c93-48a5-9589-0baaace3bb1b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.931255] env[62952]: DEBUG nova.compute.provider_tree [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.938489] env[62952]: DEBUG nova.network.neutron [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Successfully created port: 95b813f6-3248-4a6a-98b0-033734799b00 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 843.019883] env[62952]: DEBUG nova.network.neutron [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 843.047790] env[62952]: DEBUG nova.compute.manager [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 843.237180] env[62952]: DEBUG nova.network.neutron [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Updating instance_info_cache with network_info: [{"id": "3a705583-6448-4020-9894-5f2056e2e4e7", "address": "fa:16:3e:86:6e:fa", "network": {"id": "247a88e9-e8e8-46af-ac41-6777f0782fe0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1297422061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d93803f89ca64e66b8271e7b0bb48983", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a705583-64", "ovs_interfaceid": "3a705583-6448-4020-9894-5f2056e2e4e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.299211] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da8a791d-0f39-45cd-a7e3-4d0393f353cc tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "5f1eca6f-4cc4-483e-b72a-b479378277fe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.250s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.436050] env[62952]: DEBUG nova.scheduler.client.report [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 843.739080] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Releasing lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.739490] env[62952]: DEBUG nova.compute.manager [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Instance network_info: |[{"id": "3a705583-6448-4020-9894-5f2056e2e4e7", "address": "fa:16:3e:86:6e:fa", "network": {"id": "247a88e9-e8e8-46af-ac41-6777f0782fe0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1297422061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d93803f89ca64e66b8271e7b0bb48983", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a705583-64", "ovs_interfaceid": "3a705583-6448-4020-9894-5f2056e2e4e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 843.739958] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:6e:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4fcde7-8926-402a-a9b7-4878d2bc1cf6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a705583-6448-4020-9894-5f2056e2e4e7', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.747463] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Creating folder: Project (d93803f89ca64e66b8271e7b0bb48983). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 843.747747] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f8201289-e9d8-449e-8315-2e00785cfae4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.759634] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Created folder: Project (d93803f89ca64e66b8271e7b0bb48983) in parent group-v271811. [ 843.759800] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Creating folder: Instances. Parent ref: group-v271840. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 843.760241] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7042cde9-ebcc-42da-831f-e8b14057dbd2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.768044] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Created folder: Instances in parent group-v271840. [ 843.768262] env[62952]: DEBUG oslo.service.loopingcall [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.768436] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 843.768632] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d592b2bb-531c-4723-8426-03670f5e4724 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.786714] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.786714] env[62952]: value = "task-1263159" [ 843.786714] env[62952]: _type = "Task" [ 843.786714] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.794855] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263159, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.801806] env[62952]: DEBUG nova.compute.manager [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.940549] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.903s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.941316] env[62952]: ERROR nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c5ae85f8-f1b8-497a-a7b9-a9b53a3c813c, please check neutron logs for more information. [ 843.941316] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Traceback (most recent call last): [ 843.941316] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 843.941316] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] self.driver.spawn(context, instance, image_meta, [ 843.941316] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 843.941316] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 843.941316] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 843.941316] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] vm_ref = self.build_virtual_machine(instance, [ 843.941316] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 843.941316] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] vif_infos = vmwarevif.get_vif_info(self._session, [ 843.941316] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 843.941648] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] for vif in network_info: [ 843.941648] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 843.941648] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] return self._sync_wrapper(fn, *args, **kwargs) [ 843.941648] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 843.941648] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] self.wait() [ 843.941648] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 843.941648] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] self[:] = self._gt.wait() [ 843.941648] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 843.941648] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] return self._exit_event.wait() [ 843.941648] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 843.941648] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] result = hub.switch() [ 843.941648] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 843.941648] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] return self.greenlet.switch() [ 843.942230] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 843.942230] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] result = function(*args, **kwargs) [ 843.942230] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 843.942230] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] return func(*args, **kwargs) [ 843.942230] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 843.942230] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] raise e [ 843.942230] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 843.942230] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] nwinfo = self.network_api.allocate_for_instance( [ 843.942230] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 843.942230] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] created_port_ids = self._update_ports_for_instance( [ 843.942230] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 843.942230] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] with excutils.save_and_reraise_exception(): [ 843.942230] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 843.942896] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] self.force_reraise() [ 843.942896] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 843.942896] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] raise self.value [ 843.942896] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 843.942896] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] updated_port = self._update_port( [ 843.942896] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 843.942896] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] _ensure_no_port_binding_failure(port) [ 843.942896] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 843.942896] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] raise exception.PortBindingFailed(port_id=port['id']) [ 843.942896] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] nova.exception.PortBindingFailed: Binding failed for port c5ae85f8-f1b8-497a-a7b9-a9b53a3c813c, please check neutron logs for more information. [ 843.942896] env[62952]: ERROR nova.compute.manager [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] [ 843.943885] env[62952]: DEBUG nova.compute.utils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Binding failed for port c5ae85f8-f1b8-497a-a7b9-a9b53a3c813c, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 843.943885] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.605s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.947237] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Build of instance 9c5167eb-3a15-47ff-923e-c83cc3529cc8 was re-scheduled: Binding failed for port c5ae85f8-f1b8-497a-a7b9-a9b53a3c813c, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 843.947461] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 843.947728] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "refresh_cache-9c5167eb-3a15-47ff-923e-c83cc3529cc8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.947940] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquired lock "refresh_cache-9c5167eb-3a15-47ff-923e-c83cc3529cc8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.948215] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 844.059069] env[62952]: DEBUG nova.compute.manager [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 844.106750] env[62952]: DEBUG nova.virt.hardware [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 844.107284] env[62952]: DEBUG nova.virt.hardware [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 844.107615] env[62952]: DEBUG nova.virt.hardware [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.107947] env[62952]: DEBUG nova.virt.hardware [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 844.108259] env[62952]: DEBUG nova.virt.hardware [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.110017] env[62952]: DEBUG nova.virt.hardware [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 844.110017] env[62952]: DEBUG nova.virt.hardware [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 844.110017] env[62952]: DEBUG nova.virt.hardware [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 844.110017] env[62952]: DEBUG nova.virt.hardware [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 844.110017] env[62952]: DEBUG nova.virt.hardware [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 844.110503] env[62952]: DEBUG nova.virt.hardware [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.113909] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d05e7f6f-5ad1-4462-82f1-b9ff2f0d1c20 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.124244] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecdee99b-e5bb-4a6a-bdce-4c978efc8b13 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.146919] env[62952]: DEBUG nova.compute.manager [req-a0e44bf6-de25-4067-8732-eedbfecf0c3f req-ddc2fb93-7a59-46bc-91f3-7206d678424d service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Received event network-changed-3a705583-6448-4020-9894-5f2056e2e4e7 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 844.147138] env[62952]: DEBUG nova.compute.manager [req-a0e44bf6-de25-4067-8732-eedbfecf0c3f req-ddc2fb93-7a59-46bc-91f3-7206d678424d service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Refreshing instance network info cache due to event network-changed-3a705583-6448-4020-9894-5f2056e2e4e7. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 844.147896] env[62952]: DEBUG oslo_concurrency.lockutils [req-a0e44bf6-de25-4067-8732-eedbfecf0c3f req-ddc2fb93-7a59-46bc-91f3-7206d678424d service nova] Acquiring lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.147896] env[62952]: DEBUG oslo_concurrency.lockutils [req-a0e44bf6-de25-4067-8732-eedbfecf0c3f req-ddc2fb93-7a59-46bc-91f3-7206d678424d service nova] Acquired lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.147896] env[62952]: DEBUG nova.network.neutron [req-a0e44bf6-de25-4067-8732-eedbfecf0c3f req-ddc2fb93-7a59-46bc-91f3-7206d678424d service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Refreshing network info cache for port 3a705583-6448-4020-9894-5f2056e2e4e7 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 844.261465] env[62952]: DEBUG nova.compute.manager [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 844.262546] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56a208c4-b2f0-42c6-b112-2cd386af86d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.297517] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263159, 'name': CreateVM_Task, 'duration_secs': 0.382535} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.298245] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 844.298375] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.298626] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.298897] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 844.299147] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73f6fe57-6824-4dd4-871c-48ca985e9c52 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.303569] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 844.303569] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c091eb-2153-1bed-4284-26f143362903" [ 844.303569] env[62952]: _type = "Task" [ 844.303569] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.316671] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c091eb-2153-1bed-4284-26f143362903, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.335262] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.470285] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 844.712881] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8fb88e5-7590-4827-b36a-b78b36bd8199 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.721531] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7f1847-3329-4c13-ab16-345ce9e36fb3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.757330] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.759336] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0eae0c0-b025-40fb-8203-9acd47ea6a6f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.767694] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a43db97b-f3fa-4724-9343-75212b231594 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.774896] env[62952]: INFO nova.compute.manager [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] instance snapshotting [ 844.784613] env[62952]: DEBUG nova.compute.provider_tree [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.786572] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd4ea86-bf85-4107-95ff-600403968af4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.811751] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd725baf-1ba5-4543-848f-c9b27c33e5be {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.812546] env[62952]: DEBUG nova.network.neutron [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Successfully updated port: 95b813f6-3248-4a6a-98b0-033734799b00 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 844.819618] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c091eb-2153-1bed-4284-26f143362903, 'name': SearchDatastore_Task, 'duration_secs': 0.031678} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.821586] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.821943] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.822108] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.822255] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.822478] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.827185] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ae0cdb2f-bfca-447d-b21a-002287547d6a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.837411] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.837598] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 844.838733] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0881fa21-d88f-4a80-98a7-b8b845105a99 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.844211] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 844.844211] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52eae538-c0c2-280c-d913-bf13815bb38f" [ 844.844211] env[62952]: _type = "Task" [ 844.844211] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.854254] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52eae538-c0c2-280c-d913-bf13815bb38f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.968287] env[62952]: DEBUG nova.network.neutron [req-a0e44bf6-de25-4067-8732-eedbfecf0c3f req-ddc2fb93-7a59-46bc-91f3-7206d678424d service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Updated VIF entry in instance network info cache for port 3a705583-6448-4020-9894-5f2056e2e4e7. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 844.968936] env[62952]: DEBUG nova.network.neutron [req-a0e44bf6-de25-4067-8732-eedbfecf0c3f req-ddc2fb93-7a59-46bc-91f3-7206d678424d service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Updating instance_info_cache with network_info: [{"id": "3a705583-6448-4020-9894-5f2056e2e4e7", "address": "fa:16:3e:86:6e:fa", "network": {"id": "247a88e9-e8e8-46af-ac41-6777f0782fe0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1297422061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d93803f89ca64e66b8271e7b0bb48983", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a705583-64", "ovs_interfaceid": "3a705583-6448-4020-9894-5f2056e2e4e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.263437] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Releasing lock "refresh_cache-9c5167eb-3a15-47ff-923e-c83cc3529cc8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.263765] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 845.263871] env[62952]: DEBUG nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 845.264053] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 845.280615] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 845.290859] env[62952]: DEBUG nova.scheduler.client.report [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.315559] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Acquiring lock "refresh_cache-219cfb14-ca73-45d3-86e8-ed4642d7d480" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.315710] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Acquired lock "refresh_cache-219cfb14-ca73-45d3-86e8-ed4642d7d480" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.315861] env[62952]: DEBUG nova.network.neutron [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 845.331061] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 845.331570] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6f887db0-7255-410b-b131-6ef458fcc23d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.338513] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 845.338513] env[62952]: value = "task-1263160" [ 845.338513] env[62952]: _type = "Task" [ 845.338513] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.346715] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263160, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.355161] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52eae538-c0c2-280c-d913-bf13815bb38f, 'name': SearchDatastore_Task, 'duration_secs': 0.012152} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.355938] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59f51631-cd59-47d6-978b-fe49cd41072d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.360612] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 845.360612] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5284bf6c-9633-44a3-760e-90bf4c78a032" [ 845.360612] env[62952]: _type = "Task" [ 845.360612] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.369699] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5284bf6c-9633-44a3-760e-90bf4c78a032, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.471479] env[62952]: DEBUG oslo_concurrency.lockutils [req-a0e44bf6-de25-4067-8732-eedbfecf0c3f req-ddc2fb93-7a59-46bc-91f3-7206d678424d service nova] Releasing lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.783906] env[62952]: DEBUG nova.network.neutron [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.795265] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.852s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.795880] env[62952]: ERROR nova.compute.manager [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d1ec9d86-9392-41f9-95b2-2d574a3b9ef1, please check neutron logs for more information. [ 845.795880] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Traceback (most recent call last): [ 845.795880] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 845.795880] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] self.driver.spawn(context, instance, image_meta, [ 845.795880] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 845.795880] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] self._vmops.spawn(context, instance, image_meta, injected_files, [ 845.795880] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 845.795880] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] vm_ref = self.build_virtual_machine(instance, [ 845.795880] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 845.795880] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] vif_infos = vmwarevif.get_vif_info(self._session, [ 845.795880] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 845.796270] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] for vif in network_info: [ 845.796270] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 845.796270] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] return self._sync_wrapper(fn, *args, **kwargs) [ 845.796270] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 845.796270] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] self.wait() [ 845.796270] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 845.796270] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] self[:] = self._gt.wait() [ 845.796270] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 845.796270] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] return self._exit_event.wait() [ 845.796270] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 845.796270] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] result = hub.switch() [ 845.796270] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 845.796270] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] return self.greenlet.switch() [ 845.796690] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 845.796690] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] result = function(*args, **kwargs) [ 845.796690] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 845.796690] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] return func(*args, **kwargs) [ 845.796690] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 845.796690] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] raise e [ 845.796690] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 845.796690] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] nwinfo = self.network_api.allocate_for_instance( [ 845.796690] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 845.796690] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] created_port_ids = self._update_ports_for_instance( [ 845.796690] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 845.796690] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] with excutils.save_and_reraise_exception(): [ 845.796690] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 845.797231] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] self.force_reraise() [ 845.797231] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 845.797231] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] raise self.value [ 845.797231] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 845.797231] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] updated_port = self._update_port( [ 845.797231] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 845.797231] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] _ensure_no_port_binding_failure(port) [ 845.797231] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 845.797231] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] raise exception.PortBindingFailed(port_id=port['id']) [ 845.797231] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] nova.exception.PortBindingFailed: Binding failed for port d1ec9d86-9392-41f9-95b2-2d574a3b9ef1, please check neutron logs for more information. [ 845.797231] env[62952]: ERROR nova.compute.manager [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] [ 845.797597] env[62952]: DEBUG nova.compute.utils [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Binding failed for port d1ec9d86-9392-41f9-95b2-2d574a3b9ef1, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 845.797775] env[62952]: DEBUG oslo_concurrency.lockutils [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.819s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.799229] env[62952]: INFO nova.compute.claims [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.802062] env[62952]: DEBUG nova.compute.manager [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Build of instance 9be8f7f5-80eb-4e38-9501-e76abd7227db was re-scheduled: Binding failed for port d1ec9d86-9392-41f9-95b2-2d574a3b9ef1, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 845.802512] env[62952]: DEBUG nova.compute.manager [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 845.802735] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Acquiring lock "refresh_cache-9be8f7f5-80eb-4e38-9501-e76abd7227db" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.802876] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Acquired lock "refresh_cache-9be8f7f5-80eb-4e38-9501-e76abd7227db" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.803046] env[62952]: DEBUG nova.network.neutron [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 845.845571] env[62952]: DEBUG nova.network.neutron [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 845.850479] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263160, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.871804] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5284bf6c-9633-44a3-760e-90bf4c78a032, 'name': SearchDatastore_Task, 'duration_secs': 0.010011} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.872426] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.872426] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 9cc7bc2f-ff22-43cf-be79-ff8c6426115b/9cc7bc2f-ff22-43cf-be79-ff8c6426115b.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 845.872616] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c0efdeee-2519-46ef-9129-474a61fbbe69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.878752] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 845.878752] env[62952]: value = "task-1263161" [ 845.878752] env[62952]: _type = "Task" [ 845.878752] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.886339] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263161, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.981834] env[62952]: DEBUG nova.network.neutron [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Updating instance_info_cache with network_info: [{"id": "95b813f6-3248-4a6a-98b0-033734799b00", "address": "fa:16:3e:79:a3:10", "network": {"id": "9aaea3c9-901f-471b-9675-226f08d0e0ab", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1357570599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e698fc6474024bbca26d6b6c06082f03", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95b813f6-32", "ovs_interfaceid": "95b813f6-3248-4a6a-98b0-033734799b00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.183648] env[62952]: DEBUG nova.compute.manager [req-31462db8-adf0-46ab-b672-99a7459d7c10 req-f9ece3b5-aca0-46db-b6a1-47f5af045812 service nova] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Received event network-vif-plugged-95b813f6-3248-4a6a-98b0-033734799b00 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 846.184015] env[62952]: DEBUG oslo_concurrency.lockutils [req-31462db8-adf0-46ab-b672-99a7459d7c10 req-f9ece3b5-aca0-46db-b6a1-47f5af045812 service nova] Acquiring lock "219cfb14-ca73-45d3-86e8-ed4642d7d480-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.184250] env[62952]: DEBUG oslo_concurrency.lockutils [req-31462db8-adf0-46ab-b672-99a7459d7c10 req-f9ece3b5-aca0-46db-b6a1-47f5af045812 service nova] Lock "219cfb14-ca73-45d3-86e8-ed4642d7d480-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.184420] env[62952]: DEBUG oslo_concurrency.lockutils [req-31462db8-adf0-46ab-b672-99a7459d7c10 req-f9ece3b5-aca0-46db-b6a1-47f5af045812 service nova] Lock "219cfb14-ca73-45d3-86e8-ed4642d7d480-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.184660] env[62952]: DEBUG nova.compute.manager [req-31462db8-adf0-46ab-b672-99a7459d7c10 req-f9ece3b5-aca0-46db-b6a1-47f5af045812 service nova] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] No waiting events found dispatching network-vif-plugged-95b813f6-3248-4a6a-98b0-033734799b00 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 846.184851] env[62952]: WARNING nova.compute.manager [req-31462db8-adf0-46ab-b672-99a7459d7c10 req-f9ece3b5-aca0-46db-b6a1-47f5af045812 service nova] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Received unexpected event network-vif-plugged-95b813f6-3248-4a6a-98b0-033734799b00 for instance with vm_state building and task_state spawning. [ 846.185028] env[62952]: DEBUG nova.compute.manager [req-31462db8-adf0-46ab-b672-99a7459d7c10 req-f9ece3b5-aca0-46db-b6a1-47f5af045812 service nova] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Received event network-changed-95b813f6-3248-4a6a-98b0-033734799b00 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 846.185267] env[62952]: DEBUG nova.compute.manager [req-31462db8-adf0-46ab-b672-99a7459d7c10 req-f9ece3b5-aca0-46db-b6a1-47f5af045812 service nova] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Refreshing instance network info cache due to event network-changed-95b813f6-3248-4a6a-98b0-033734799b00. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 846.185456] env[62952]: DEBUG oslo_concurrency.lockutils [req-31462db8-adf0-46ab-b672-99a7459d7c10 req-f9ece3b5-aca0-46db-b6a1-47f5af045812 service nova] Acquiring lock "refresh_cache-219cfb14-ca73-45d3-86e8-ed4642d7d480" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.286844] env[62952]: INFO nova.compute.manager [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 9c5167eb-3a15-47ff-923e-c83cc3529cc8] Took 1.02 seconds to deallocate network for instance. [ 846.324031] env[62952]: DEBUG nova.network.neutron [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 846.351840] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263160, 'name': CreateSnapshot_Task, 'duration_secs': 0.824702} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.352152] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 846.352926] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b665f3cc-2a25-4e38-8919-414868223d8b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.387650] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263161, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.402160] env[62952]: DEBUG nova.network.neutron [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.485055] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Releasing lock "refresh_cache-219cfb14-ca73-45d3-86e8-ed4642d7d480" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.485405] env[62952]: DEBUG nova.compute.manager [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Instance network_info: |[{"id": "95b813f6-3248-4a6a-98b0-033734799b00", "address": "fa:16:3e:79:a3:10", "network": {"id": "9aaea3c9-901f-471b-9675-226f08d0e0ab", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1357570599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e698fc6474024bbca26d6b6c06082f03", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95b813f6-32", "ovs_interfaceid": "95b813f6-3248-4a6a-98b0-033734799b00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 846.485702] env[62952]: DEBUG oslo_concurrency.lockutils [req-31462db8-adf0-46ab-b672-99a7459d7c10 req-f9ece3b5-aca0-46db-b6a1-47f5af045812 service nova] Acquired lock "refresh_cache-219cfb14-ca73-45d3-86e8-ed4642d7d480" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.485878] env[62952]: DEBUG nova.network.neutron [req-31462db8-adf0-46ab-b672-99a7459d7c10 req-f9ece3b5-aca0-46db-b6a1-47f5af045812 service nova] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Refreshing network info cache for port 95b813f6-3248-4a6a-98b0-033734799b00 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 846.487575] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:79:a3:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '95b813f6-3248-4a6a-98b0-033734799b00', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 846.494442] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Creating folder: Project (e698fc6474024bbca26d6b6c06082f03). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 846.497188] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8afe30bc-a0f8-4f22-99b4-7b9fca58418c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.507797] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Created folder: Project (e698fc6474024bbca26d6b6c06082f03) in parent group-v271811. [ 846.507984] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Creating folder: Instances. Parent ref: group-v271844. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 846.509120] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7db20354-a850-4a07-81d6-bb780461913b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.516255] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Created folder: Instances in parent group-v271844. [ 846.516483] env[62952]: DEBUG oslo.service.loopingcall [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.516664] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 846.516857] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5dbaf29c-c36a-4460-a390-c21bff72593c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.539140] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 846.539140] env[62952]: value = "task-1263164" [ 846.539140] env[62952]: _type = "Task" [ 846.539140] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.546327] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263164, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.700459] env[62952]: DEBUG nova.network.neutron [req-31462db8-adf0-46ab-b672-99a7459d7c10 req-f9ece3b5-aca0-46db-b6a1-47f5af045812 service nova] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Updated VIF entry in instance network info cache for port 95b813f6-3248-4a6a-98b0-033734799b00. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 846.701107] env[62952]: DEBUG nova.network.neutron [req-31462db8-adf0-46ab-b672-99a7459d7c10 req-f9ece3b5-aca0-46db-b6a1-47f5af045812 service nova] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Updating instance_info_cache with network_info: [{"id": "95b813f6-3248-4a6a-98b0-033734799b00", "address": "fa:16:3e:79:a3:10", "network": {"id": "9aaea3c9-901f-471b-9675-226f08d0e0ab", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1357570599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e698fc6474024bbca26d6b6c06082f03", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95b813f6-32", "ovs_interfaceid": "95b813f6-3248-4a6a-98b0-033734799b00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.870761] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 847.006283] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-57cb1001-f7c7-47ec-8447-90eaf2c26fa3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.006283] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 847.006283] env[62952]: value = "task-1263165" [ 847.006283] env[62952]: _type = "Task" [ 847.006283] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.006283] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Releasing lock "refresh_cache-9be8f7f5-80eb-4e38-9501-e76abd7227db" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.006283] env[62952]: DEBUG nova.compute.manager [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 847.006283] env[62952]: DEBUG nova.compute.manager [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 847.006988] env[62952]: DEBUG nova.network.neutron [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 847.006988] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263161, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531347} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.006988] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 9cc7bc2f-ff22-43cf-be79-ff8c6426115b/9cc7bc2f-ff22-43cf-be79-ff8c6426115b.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 847.006988] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 847.006988] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fcc59029-221e-43d3-9080-42d5fae28210 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.007217] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 847.007217] env[62952]: value = "task-1263166" [ 847.007217] env[62952]: _type = "Task" [ 847.007217] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.010727] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263165, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.014548] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263166, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.023930] env[62952]: DEBUG nova.network.neutron [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 847.055822] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263164, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.204068] env[62952]: DEBUG oslo_concurrency.lockutils [req-31462db8-adf0-46ab-b672-99a7459d7c10 req-f9ece3b5-aca0-46db-b6a1-47f5af045812 service nova] Releasing lock "refresh_cache-219cfb14-ca73-45d3-86e8-ed4642d7d480" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.310987] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd7e385-fdbb-43d7-81b8-0d013ef6d6d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.319062] env[62952]: INFO nova.scheduler.client.report [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Deleted allocations for instance 9c5167eb-3a15-47ff-923e-c83cc3529cc8 [ 847.326075] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3184dd7a-ef06-435b-a117-1eca3c39ee1f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.357988] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3675b6-035d-487d-8ddc-bdf131b59de0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.367930] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19906386-3789-4138-95ce-9be35709ff7f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.115968] env[62952]: DEBUG nova.network.neutron [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.117372] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6b6127f-c508-4dae-8ecc-6edc86155d60 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "9c5167eb-3a15-47ff-923e-c83cc3529cc8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.428s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.117805] env[62952]: DEBUG nova.compute.provider_tree [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.137277] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263166, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077355} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.140500] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 848.141111] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263165, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.141831] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-774338c0-cb35-4c8f-b1f7-525a87b7b5a7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.149759] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263164, 'name': CreateVM_Task, 'duration_secs': 0.559297} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.158884] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 848.168420] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] 9cc7bc2f-ff22-43cf-be79-ff8c6426115b/9cc7bc2f-ff22-43cf-be79-ff8c6426115b.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.169135] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.169281] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.169605] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 848.170089] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea5fcbaf-f493-4862-9707-ca99d7219fa0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.185249] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2865bc58-04f8-4ec6-8dc9-5185a20c3624 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.190633] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Waiting for the task: (returnval){ [ 848.190633] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52172672-9b57-0c61-4fe6-cd5cac878ecc" [ 848.190633] env[62952]: _type = "Task" [ 848.190633] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.194733] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 848.194733] env[62952]: value = "task-1263167" [ 848.194733] env[62952]: _type = "Task" [ 848.194733] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.201466] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52172672-9b57-0c61-4fe6-cd5cac878ecc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.206501] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263167, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.623040] env[62952]: INFO nova.compute.manager [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] [instance: 9be8f7f5-80eb-4e38-9501-e76abd7227db] Took 1.71 seconds to deallocate network for instance. [ 848.629826] env[62952]: DEBUG nova.scheduler.client.report [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.632975] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263165, 'name': CloneVM_Task} progress is 95%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.633297] env[62952]: DEBUG nova.compute.manager [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 848.702941] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52172672-9b57-0c61-4fe6-cd5cac878ecc, 'name': SearchDatastore_Task, 'duration_secs': 0.009956} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.703491] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.703752] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 848.703982] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.704171] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.704369] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 848.704617] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-83ea74fc-5f14-4e7b-8583-860890cec42c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.709226] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263167, 'name': ReconfigVM_Task, 'duration_secs': 0.268255} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.709788] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Reconfigured VM instance instance-00000033 to attach disk [datastore1] 9cc7bc2f-ff22-43cf-be79-ff8c6426115b/9cc7bc2f-ff22-43cf-be79-ff8c6426115b.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.710383] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0882b394-47cc-4b45-813a-4f7f9fcac1ea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.715435] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 848.715604] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 848.717171] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01479532-aabe-4b29-a5c8-ea211c228c74 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.719507] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 848.719507] env[62952]: value = "task-1263168" [ 848.719507] env[62952]: _type = "Task" [ 848.719507] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.726848] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Waiting for the task: (returnval){ [ 848.726848] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52bef0e8-3301-955c-6c5c-3dc55298bb80" [ 848.726848] env[62952]: _type = "Task" [ 848.726848] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.733734] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263168, 'name': Rename_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.739774] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52bef0e8-3301-955c-6c5c-3dc55298bb80, 'name': SearchDatastore_Task, 'duration_secs': 0.009256} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.740720] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-449fea17-812a-4952-9ba6-77c598a87d19 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.746561] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Waiting for the task: (returnval){ [ 848.746561] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526f0403-f14b-3558-6058-b825f106038e" [ 848.746561] env[62952]: _type = "Task" [ 848.746561] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.755448] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526f0403-f14b-3558-6058-b825f106038e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.129232] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263165, 'name': CloneVM_Task, 'duration_secs': 2.064461} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.130190] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Created linked-clone VM from snapshot [ 849.130486] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59940be1-29f3-49c1-a04f-66917b1ec7f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.138616] env[62952]: DEBUG oslo_concurrency.lockutils [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.341s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.139149] env[62952]: DEBUG nova.compute.manager [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 849.143638] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Uploading image 8518bf0a-b7da-4699-a0d0-32c8ad4958d8 {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 849.148031] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.399s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.148850] env[62952]: INFO nova.compute.claims [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 849.161621] env[62952]: DEBUG oslo_concurrency.lockutils [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.171517] env[62952]: DEBUG oslo_vmware.rw_handles [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 849.171517] env[62952]: value = "vm-271847" [ 849.171517] env[62952]: _type = "VirtualMachine" [ 849.171517] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 849.172062] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-03b4f6f3-1db3-4968-95bd-d3fbac518178 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.179559] env[62952]: DEBUG oslo_vmware.rw_handles [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lease: (returnval){ [ 849.179559] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52cabb94-577b-a6a1-850c-d9da8311597e" [ 849.179559] env[62952]: _type = "HttpNfcLease" [ 849.179559] env[62952]: } obtained for exporting VM: (result){ [ 849.179559] env[62952]: value = "vm-271847" [ 849.179559] env[62952]: _type = "VirtualMachine" [ 849.179559] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 849.179812] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the lease: (returnval){ [ 849.179812] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52cabb94-577b-a6a1-850c-d9da8311597e" [ 849.179812] env[62952]: _type = "HttpNfcLease" [ 849.179812] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 849.186506] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 849.186506] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52cabb94-577b-a6a1-850c-d9da8311597e" [ 849.186506] env[62952]: _type = "HttpNfcLease" [ 849.186506] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 849.229167] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263168, 'name': Rename_Task, 'duration_secs': 0.138043} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.229459] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 849.229698] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b409a156-3031-42de-a94c-fc2552bcb985 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.235092] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 849.235092] env[62952]: value = "task-1263170" [ 849.235092] env[62952]: _type = "Task" [ 849.235092] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.242064] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263170, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.254469] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526f0403-f14b-3558-6058-b825f106038e, 'name': SearchDatastore_Task, 'duration_secs': 0.009455} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.254690] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.254933] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 219cfb14-ca73-45d3-86e8-ed4642d7d480/219cfb14-ca73-45d3-86e8-ed4642d7d480.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 849.255215] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6036aae5-5414-4358-b841-87e17133997d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.261343] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Waiting for the task: (returnval){ [ 849.261343] env[62952]: value = "task-1263171" [ 849.261343] env[62952]: _type = "Task" [ 849.261343] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.268092] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': task-1263171, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.429714] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "e5e7b2b5-34cf-4d37-87ff-1b57264a3516" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.430041] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "e5e7b2b5-34cf-4d37-87ff-1b57264a3516" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.454972] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "051f9279-533a-4545-b9e2-272bf40956f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.455205] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "051f9279-533a-4545-b9e2-272bf40956f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.650117] env[62952]: DEBUG nova.compute.utils [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 849.652199] env[62952]: DEBUG nova.compute.manager [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Not allocating networking since 'none' was specified. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 849.657258] env[62952]: INFO nova.scheduler.client.report [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Deleted allocations for instance 9be8f7f5-80eb-4e38-9501-e76abd7227db [ 849.691115] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 849.691115] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52cabb94-577b-a6a1-850c-d9da8311597e" [ 849.691115] env[62952]: _type = "HttpNfcLease" [ 849.691115] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 849.691520] env[62952]: DEBUG oslo_vmware.rw_handles [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 849.691520] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52cabb94-577b-a6a1-850c-d9da8311597e" [ 849.691520] env[62952]: _type = "HttpNfcLease" [ 849.691520] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 849.692309] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d31a73b-4edc-46ed-be9c-d73599eca50d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.701728] env[62952]: DEBUG oslo_vmware.rw_handles [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5254cd6e-dfe7-7bde-995d-13fcf171fbc5/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 849.701925] env[62952]: DEBUG oslo_vmware.rw_handles [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5254cd6e-dfe7-7bde-995d-13fcf171fbc5/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 849.776671] env[62952]: DEBUG oslo_vmware.api [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263170, 'name': PowerOnVM_Task, 'duration_secs': 0.511343} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.776889] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': task-1263171, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463598} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.777135] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 849.777336] env[62952]: INFO nova.compute.manager [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Took 8.16 seconds to spawn the instance on the hypervisor. [ 849.777510] env[62952]: DEBUG nova.compute.manager [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.777781] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 219cfb14-ca73-45d3-86e8-ed4642d7d480/219cfb14-ca73-45d3-86e8-ed4642d7d480.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 849.777970] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 849.778670] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b82f1cf5-d933-4d4f-8267-3715f96b47ed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.780998] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-809f37a0-671a-4309-a8e8-c94e0d6e19a9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.789372] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Waiting for the task: (returnval){ [ 849.789372] env[62952]: value = "task-1263172" [ 849.789372] env[62952]: _type = "Task" [ 849.789372] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.796896] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': task-1263172, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.829902] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-66376907-5815-45d9-95b3-99ee144855a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.155614] env[62952]: DEBUG nova.compute.manager [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 850.165846] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c84dacca-8447-43ff-b564-e72ef0f37a94 tempest-ServerActionsTestOtherA-506680640 tempest-ServerActionsTestOtherA-506680640-project-member] Lock "9be8f7f5-80eb-4e38-9501-e76abd7227db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.861s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.312990] env[62952]: INFO nova.compute.manager [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Took 34.95 seconds to build instance. [ 850.320502] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': task-1263172, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069148} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.321912] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 850.322525] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5536313a-3fd1-429b-a3fc-41f60c78f041 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.352683] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 219cfb14-ca73-45d3-86e8-ed4642d7d480/219cfb14-ca73-45d3-86e8-ed4642d7d480.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 850.356338] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a6079cf-2ed0-4d28-9bd4-a750fb12cf22 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.380508] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Waiting for the task: (returnval){ [ 850.380508] env[62952]: value = "task-1263173" [ 850.380508] env[62952]: _type = "Task" [ 850.380508] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.396608] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': task-1263173, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.572912] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2428f2f5-6bce-493c-a633-cc1d3c974aa2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.581098] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a3800c-c472-4819-b575-8582722767c9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.612296] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5754f82c-f99a-479a-974d-694e848f7ef3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.620258] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-169073c0-a5a5-49c2-81f3-386e6b3cd0f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.635259] env[62952]: DEBUG nova.compute.provider_tree [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.668172] env[62952]: DEBUG nova.compute.manager [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 850.788237] env[62952]: DEBUG nova.compute.manager [req-abc223b3-d458-4f12-b611-2ddfab126163 req-1cfa8a45-1bae-46e6-856c-18e60a93851c service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Received event network-changed-3a705583-6448-4020-9894-5f2056e2e4e7 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 850.788237] env[62952]: DEBUG nova.compute.manager [req-abc223b3-d458-4f12-b611-2ddfab126163 req-1cfa8a45-1bae-46e6-856c-18e60a93851c service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Refreshing instance network info cache due to event network-changed-3a705583-6448-4020-9894-5f2056e2e4e7. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 850.788237] env[62952]: DEBUG oslo_concurrency.lockutils [req-abc223b3-d458-4f12-b611-2ddfab126163 req-1cfa8a45-1bae-46e6-856c-18e60a93851c service nova] Acquiring lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.788237] env[62952]: DEBUG oslo_concurrency.lockutils [req-abc223b3-d458-4f12-b611-2ddfab126163 req-1cfa8a45-1bae-46e6-856c-18e60a93851c service nova] Acquired lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.788237] env[62952]: DEBUG nova.network.neutron [req-abc223b3-d458-4f12-b611-2ddfab126163 req-1cfa8a45-1bae-46e6-856c-18e60a93851c service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Refreshing network info cache for port 3a705583-6448-4020-9894-5f2056e2e4e7 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 850.822127] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c5afcc9-8abe-413e-b0d7-b45bc38f3be6 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.793s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.893021] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': task-1263173, 'name': ReconfigVM_Task, 'duration_secs': 0.284141} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.893021] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 219cfb14-ca73-45d3-86e8-ed4642d7d480/219cfb14-ca73-45d3-86e8-ed4642d7d480.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.893021] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-20d6db77-c571-493d-8697-a62a90665ee5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.898716] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Waiting for the task: (returnval){ [ 850.898716] env[62952]: value = "task-1263174" [ 850.898716] env[62952]: _type = "Task" [ 850.898716] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.909331] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': task-1263174, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.141024] env[62952]: DEBUG nova.scheduler.client.report [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.168309] env[62952]: DEBUG nova.compute.manager [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 851.201692] env[62952]: DEBUG nova.virt.hardware [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 851.201959] env[62952]: DEBUG nova.virt.hardware [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 851.202164] env[62952]: DEBUG nova.virt.hardware [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 851.202367] env[62952]: DEBUG nova.virt.hardware [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 851.202523] env[62952]: DEBUG nova.virt.hardware [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 851.202879] env[62952]: DEBUG nova.virt.hardware [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 851.203116] env[62952]: DEBUG nova.virt.hardware [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 851.203293] env[62952]: DEBUG nova.virt.hardware [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 851.203505] env[62952]: DEBUG nova.virt.hardware [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 851.203679] env[62952]: DEBUG nova.virt.hardware [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 851.203857] env[62952]: DEBUG nova.virt.hardware [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 851.204748] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b125db5-efa6-4c9d-946e-2d1e8fd91e9a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.212335] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af5d0921-1ac1-4f8b-8a99-d56b8b51a0e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.226393] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 851.231947] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Creating folder: Project (7b82e6330774445ba79867b6315a0366). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 851.232285] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-63105c9d-7ff5-46e4-8b1b-422cfb67ea90 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.243689] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Created folder: Project (7b82e6330774445ba79867b6315a0366) in parent group-v271811. [ 851.243886] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Creating folder: Instances. Parent ref: group-v271848. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 851.244159] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e7c0a342-b4c2-4072-869e-00863b7af0e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.253177] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Created folder: Instances in parent group-v271848. [ 851.253177] env[62952]: DEBUG oslo.service.loopingcall [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.253280] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 851.253464] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-42dd437e-4b8a-42f7-8661-71111f9c033b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.270227] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 851.270227] env[62952]: value = "task-1263177" [ 851.270227] env[62952]: _type = "Task" [ 851.270227] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.277969] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263177, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.285765] env[62952]: DEBUG oslo_concurrency.lockutils [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.325190] env[62952]: DEBUG nova.compute.manager [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 851.410525] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': task-1263174, 'name': Rename_Task, 'duration_secs': 0.135138} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.410827] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 851.411102] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4569dcc5-9341-4748-bc29-e41aba340a4b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.417959] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Waiting for the task: (returnval){ [ 851.417959] env[62952]: value = "task-1263178" [ 851.417959] env[62952]: _type = "Task" [ 851.417959] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.428321] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': task-1263178, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.644390] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.644911] env[62952]: DEBUG nova.compute.manager [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 851.652030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.573s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.652030] env[62952]: INFO nova.compute.claims [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.684568] env[62952]: DEBUG nova.network.neutron [req-abc223b3-d458-4f12-b611-2ddfab126163 req-1cfa8a45-1bae-46e6-856c-18e60a93851c service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Updated VIF entry in instance network info cache for port 3a705583-6448-4020-9894-5f2056e2e4e7. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 851.684939] env[62952]: DEBUG nova.network.neutron [req-abc223b3-d458-4f12-b611-2ddfab126163 req-1cfa8a45-1bae-46e6-856c-18e60a93851c service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Updating instance_info_cache with network_info: [{"id": "3a705583-6448-4020-9894-5f2056e2e4e7", "address": "fa:16:3e:86:6e:fa", "network": {"id": "247a88e9-e8e8-46af-ac41-6777f0782fe0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1297422061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d93803f89ca64e66b8271e7b0bb48983", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a705583-64", "ovs_interfaceid": "3a705583-6448-4020-9894-5f2056e2e4e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.780898] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263177, 'name': CreateVM_Task, 'duration_secs': 0.380666} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.781183] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 851.781551] env[62952]: DEBUG oslo_concurrency.lockutils [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.781748] env[62952]: DEBUG oslo_concurrency.lockutils [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.782101] env[62952]: DEBUG oslo_concurrency.lockutils [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.782393] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b98b7b8-e00b-47a6-9304-9f2fa4331689 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.786969] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 851.786969] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52072437-45cb-80c1-f907-c808b117e29a" [ 851.786969] env[62952]: _type = "Task" [ 851.786969] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.794915] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52072437-45cb-80c1-f907-c808b117e29a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.848962] env[62952]: DEBUG oslo_concurrency.lockutils [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.928658] env[62952]: DEBUG oslo_vmware.api [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': task-1263178, 'name': PowerOnVM_Task, 'duration_secs': 0.435153} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.929040] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 851.929333] env[62952]: INFO nova.compute.manager [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Took 7.87 seconds to spawn the instance on the hypervisor. [ 851.929530] env[62952]: DEBUG nova.compute.manager [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.930334] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30948f3f-0f6f-4459-9bf8-f68995082bed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.156023] env[62952]: DEBUG nova.compute.utils [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 852.157501] env[62952]: DEBUG nova.compute.manager [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 852.157831] env[62952]: DEBUG nova.network.neutron [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 852.187997] env[62952]: DEBUG oslo_concurrency.lockutils [req-abc223b3-d458-4f12-b611-2ddfab126163 req-1cfa8a45-1bae-46e6-856c-18e60a93851c service nova] Releasing lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.218817] env[62952]: DEBUG nova.policy [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '42a664b22bb34b74bd6ed390e40e0b64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfecd7aa86044e6bba93d20fb7017e12', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 852.304707] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52072437-45cb-80c1-f907-c808b117e29a, 'name': SearchDatastore_Task, 'duration_secs': 0.012019} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.305128] env[62952]: DEBUG oslo_concurrency.lockutils [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.305471] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.305815] env[62952]: DEBUG oslo_concurrency.lockutils [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.306048] env[62952]: DEBUG oslo_concurrency.lockutils [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.306329] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.307115] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-12bdd112-d3c3-4dd7-8357-c7082d2845be {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.318499] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.318499] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 852.318877] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-949e1cac-0524-458d-b7fe-361bbd36ebdb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.326458] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 852.326458] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525e4ac2-29eb-85bd-d7bb-37ea6b6afe08" [ 852.326458] env[62952]: _type = "Task" [ 852.326458] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.338127] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525e4ac2-29eb-85bd-d7bb-37ea6b6afe08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.449435] env[62952]: INFO nova.compute.manager [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Took 35.26 seconds to build instance. [ 852.546206] env[62952]: DEBUG nova.network.neutron [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Successfully created port: 52b68252-6346-4743-b795-f02ea8effef1 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 852.660486] env[62952]: DEBUG nova.compute.manager [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 852.851089] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525e4ac2-29eb-85bd-d7bb-37ea6b6afe08, 'name': SearchDatastore_Task, 'duration_secs': 0.012902} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.852032] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03c24f11-726f-41c4-a7bb-c3a976568395 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.861011] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 852.861011] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52eb0897-822c-98d0-2936-87eed2cc6529" [ 852.861011] env[62952]: _type = "Task" [ 852.861011] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.873539] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52eb0897-822c-98d0-2936-87eed2cc6529, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.951815] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ac40072-9d9b-40a5-9703-f4b324f411af tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Lock "219cfb14-ca73-45d3-86e8-ed4642d7d480" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.652s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.205489] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fdbb60f-99f6-4a66-a3df-4e02518e2850 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.216558] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73ff7055-dae2-4e78-b606-fcbbaa71bf11 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.254528] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1051fd0-8046-4cfd-94d9-e0a2ef708f2a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.263789] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-763ed02b-8116-43ab-a3a6-2de097801ecb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.279703] env[62952]: DEBUG nova.compute.provider_tree [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.342375] env[62952]: DEBUG nova.compute.manager [req-fe9eb48d-8255-4ab1-8fbd-6a8404eaa672 req-ff22bd6d-0232-4f12-8c23-57b9937a26c2 service nova] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Received event network-changed-95b813f6-3248-4a6a-98b0-033734799b00 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 853.342582] env[62952]: DEBUG nova.compute.manager [req-fe9eb48d-8255-4ab1-8fbd-6a8404eaa672 req-ff22bd6d-0232-4f12-8c23-57b9937a26c2 service nova] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Refreshing instance network info cache due to event network-changed-95b813f6-3248-4a6a-98b0-033734799b00. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 853.342832] env[62952]: DEBUG oslo_concurrency.lockutils [req-fe9eb48d-8255-4ab1-8fbd-6a8404eaa672 req-ff22bd6d-0232-4f12-8c23-57b9937a26c2 service nova] Acquiring lock "refresh_cache-219cfb14-ca73-45d3-86e8-ed4642d7d480" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.342984] env[62952]: DEBUG oslo_concurrency.lockutils [req-fe9eb48d-8255-4ab1-8fbd-6a8404eaa672 req-ff22bd6d-0232-4f12-8c23-57b9937a26c2 service nova] Acquired lock "refresh_cache-219cfb14-ca73-45d3-86e8-ed4642d7d480" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.343521] env[62952]: DEBUG nova.network.neutron [req-fe9eb48d-8255-4ab1-8fbd-6a8404eaa672 req-ff22bd6d-0232-4f12-8c23-57b9937a26c2 service nova] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Refreshing network info cache for port 95b813f6-3248-4a6a-98b0-033734799b00 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 853.372887] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52eb0897-822c-98d0-2936-87eed2cc6529, 'name': SearchDatastore_Task, 'duration_secs': 0.013262} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.373188] env[62952]: DEBUG oslo_concurrency.lockutils [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.373537] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] bc369b94-2f40-444e-88d7-069fd33e2646/bc369b94-2f40-444e-88d7-069fd33e2646.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 853.373829] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1db2e6a2-f706-4f28-b77c-d4a6b2b8bb86 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.382559] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 853.382559] env[62952]: value = "task-1263179" [ 853.382559] env[62952]: _type = "Task" [ 853.382559] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.392525] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263179, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.455962] env[62952]: DEBUG nova.compute.manager [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 853.670240] env[62952]: DEBUG nova.compute.manager [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 853.693969] env[62952]: DEBUG nova.virt.hardware [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 853.694274] env[62952]: DEBUG nova.virt.hardware [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 853.694457] env[62952]: DEBUG nova.virt.hardware [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 853.694660] env[62952]: DEBUG nova.virt.hardware [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 853.694825] env[62952]: DEBUG nova.virt.hardware [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 853.695008] env[62952]: DEBUG nova.virt.hardware [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 853.695572] env[62952]: DEBUG nova.virt.hardware [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 853.695850] env[62952]: DEBUG nova.virt.hardware [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 853.696114] env[62952]: DEBUG nova.virt.hardware [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 853.696506] env[62952]: DEBUG nova.virt.hardware [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 853.696787] env[62952]: DEBUG nova.virt.hardware [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 853.697784] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f03223-b952-4ff8-8c2c-739bbf467cd5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.707135] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21288bc8-5e80-48bc-936f-8a8d1cf8fadc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.783981] env[62952]: DEBUG nova.scheduler.client.report [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 853.895826] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263179, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.024766] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.169497] env[62952]: DEBUG nova.compute.manager [req-03472979-7bce-44dc-9333-57b9df6873dc req-afde391b-3c01-486e-a73e-5e84e99cc676 service nova] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Received event network-vif-plugged-52b68252-6346-4743-b795-f02ea8effef1 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 854.169782] env[62952]: DEBUG oslo_concurrency.lockutils [req-03472979-7bce-44dc-9333-57b9df6873dc req-afde391b-3c01-486e-a73e-5e84e99cc676 service nova] Acquiring lock "c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.170381] env[62952]: DEBUG oslo_concurrency.lockutils [req-03472979-7bce-44dc-9333-57b9df6873dc req-afde391b-3c01-486e-a73e-5e84e99cc676 service nova] Lock "c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.170616] env[62952]: DEBUG oslo_concurrency.lockutils [req-03472979-7bce-44dc-9333-57b9df6873dc req-afde391b-3c01-486e-a73e-5e84e99cc676 service nova] Lock "c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.170935] env[62952]: DEBUG nova.compute.manager [req-03472979-7bce-44dc-9333-57b9df6873dc req-afde391b-3c01-486e-a73e-5e84e99cc676 service nova] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] No waiting events found dispatching network-vif-plugged-52b68252-6346-4743-b795-f02ea8effef1 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 854.171026] env[62952]: WARNING nova.compute.manager [req-03472979-7bce-44dc-9333-57b9df6873dc req-afde391b-3c01-486e-a73e-5e84e99cc676 service nova] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Received unexpected event network-vif-plugged-52b68252-6346-4743-b795-f02ea8effef1 for instance with vm_state building and task_state spawning. [ 854.247261] env[62952]: DEBUG nova.network.neutron [req-fe9eb48d-8255-4ab1-8fbd-6a8404eaa672 req-ff22bd6d-0232-4f12-8c23-57b9937a26c2 service nova] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Updated VIF entry in instance network info cache for port 95b813f6-3248-4a6a-98b0-033734799b00. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 854.247628] env[62952]: DEBUG nova.network.neutron [req-fe9eb48d-8255-4ab1-8fbd-6a8404eaa672 req-ff22bd6d-0232-4f12-8c23-57b9937a26c2 service nova] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Updating instance_info_cache with network_info: [{"id": "95b813f6-3248-4a6a-98b0-033734799b00", "address": "fa:16:3e:79:a3:10", "network": {"id": "9aaea3c9-901f-471b-9675-226f08d0e0ab", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1357570599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e698fc6474024bbca26d6b6c06082f03", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95b813f6-32", "ovs_interfaceid": "95b813f6-3248-4a6a-98b0-033734799b00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.290709] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.643s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.291338] env[62952]: DEBUG nova.compute.manager [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 854.293949] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.355s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.295785] env[62952]: INFO nova.compute.claims [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 854.394814] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263179, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.646045} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.395123] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] bc369b94-2f40-444e-88d7-069fd33e2646/bc369b94-2f40-444e-88d7-069fd33e2646.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 854.395405] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.395684] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d1725e23-c105-4f63-9fe0-546841ea5af2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.406823] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 854.406823] env[62952]: value = "task-1263180" [ 854.406823] env[62952]: _type = "Task" [ 854.406823] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.417464] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263180, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.743973] env[62952]: DEBUG nova.network.neutron [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Successfully updated port: 52b68252-6346-4743-b795-f02ea8effef1 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 854.750408] env[62952]: DEBUG oslo_concurrency.lockutils [req-fe9eb48d-8255-4ab1-8fbd-6a8404eaa672 req-ff22bd6d-0232-4f12-8c23-57b9937a26c2 service nova] Releasing lock "refresh_cache-219cfb14-ca73-45d3-86e8-ed4642d7d480" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.773552] env[62952]: DEBUG nova.compute.manager [req-6f046705-1537-49d2-97e6-ff2169a06a0a req-ba881308-a31b-42e3-b5a6-67c839f3dbf4 service nova] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Received event network-changed-52b68252-6346-4743-b795-f02ea8effef1 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 854.773839] env[62952]: DEBUG nova.compute.manager [req-6f046705-1537-49d2-97e6-ff2169a06a0a req-ba881308-a31b-42e3-b5a6-67c839f3dbf4 service nova] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Refreshing instance network info cache due to event network-changed-52b68252-6346-4743-b795-f02ea8effef1. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 854.774271] env[62952]: DEBUG oslo_concurrency.lockutils [req-6f046705-1537-49d2-97e6-ff2169a06a0a req-ba881308-a31b-42e3-b5a6-67c839f3dbf4 service nova] Acquiring lock "refresh_cache-c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.774480] env[62952]: DEBUG oslo_concurrency.lockutils [req-6f046705-1537-49d2-97e6-ff2169a06a0a req-ba881308-a31b-42e3-b5a6-67c839f3dbf4 service nova] Acquired lock "refresh_cache-c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.774767] env[62952]: DEBUG nova.network.neutron [req-6f046705-1537-49d2-97e6-ff2169a06a0a req-ba881308-a31b-42e3-b5a6-67c839f3dbf4 service nova] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Refreshing network info cache for port 52b68252-6346-4743-b795-f02ea8effef1 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 854.801953] env[62952]: DEBUG nova.compute.utils [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.804887] env[62952]: DEBUG nova.compute.manager [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 854.805098] env[62952]: DEBUG nova.network.neutron [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 854.864129] env[62952]: DEBUG nova.policy [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '42a664b22bb34b74bd6ed390e40e0b64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfecd7aa86044e6bba93d20fb7017e12', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 854.917694] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263180, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081922} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.918031] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.918875] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8adebee1-d210-4181-8655-38816b51c992 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.940964] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] bc369b94-2f40-444e-88d7-069fd33e2646/bc369b94-2f40-444e-88d7-069fd33e2646.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.941318] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-251be613-4579-4b87-bbb6-c5613363918d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.963995] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 854.963995] env[62952]: value = "task-1263181" [ 854.963995] env[62952]: _type = "Task" [ 854.963995] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.973325] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263181, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.155071] env[62952]: DEBUG nova.network.neutron [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Successfully created port: ef78c1ce-2d7f-4e00-87a2-95a93cfc3f37 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 855.247007] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "refresh_cache-c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.305414] env[62952]: DEBUG nova.compute.manager [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 855.328052] env[62952]: DEBUG nova.network.neutron [req-6f046705-1537-49d2-97e6-ff2169a06a0a req-ba881308-a31b-42e3-b5a6-67c839f3dbf4 service nova] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 855.451207] env[62952]: DEBUG nova.network.neutron [req-6f046705-1537-49d2-97e6-ff2169a06a0a req-ba881308-a31b-42e3-b5a6-67c839f3dbf4 service nova] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.475635] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263181, 'name': ReconfigVM_Task, 'duration_secs': 0.391284} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.478508] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Reconfigured VM instance instance-00000035 to attach disk [datastore2] bc369b94-2f40-444e-88d7-069fd33e2646/bc369b94-2f40-444e-88d7-069fd33e2646.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.479537] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6cba3cde-68f0-469c-809b-a5a207db10ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.487942] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 855.487942] env[62952]: value = "task-1263182" [ 855.487942] env[62952]: _type = "Task" [ 855.487942] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.500147] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263182, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.657920] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-996f51f3-8bbc-4aa7-9661-fc95a6d12f6f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.667087] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f65eaa-68d0-487b-882e-3eb0a1555f97 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.699097] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4823d086-ee27-40c7-991e-d65f691b4db2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.709292] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a95a94-e720-46c3-96a8-a023f7414410 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.725620] env[62952]: DEBUG nova.compute.provider_tree [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.954407] env[62952]: DEBUG oslo_concurrency.lockutils [req-6f046705-1537-49d2-97e6-ff2169a06a0a req-ba881308-a31b-42e3-b5a6-67c839f3dbf4 service nova] Releasing lock "refresh_cache-c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.954912] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquired lock "refresh_cache-c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.955098] env[62952]: DEBUG nova.network.neutron [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 856.006383] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263182, 'name': Rename_Task, 'duration_secs': 0.179658} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.006766] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 856.007096] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-463cfde3-ec53-4e31-895c-47701411cfcd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.016031] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 856.016031] env[62952]: value = "task-1263183" [ 856.016031] env[62952]: _type = "Task" [ 856.016031] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.026991] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263183, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.229626] env[62952]: DEBUG nova.scheduler.client.report [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.318104] env[62952]: DEBUG nova.compute.manager [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 856.341651] env[62952]: DEBUG nova.virt.hardware [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 856.341950] env[62952]: DEBUG nova.virt.hardware [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 856.342179] env[62952]: DEBUG nova.virt.hardware [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.342480] env[62952]: DEBUG nova.virt.hardware [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 856.342668] env[62952]: DEBUG nova.virt.hardware [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.342863] env[62952]: DEBUG nova.virt.hardware [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 856.343153] env[62952]: DEBUG nova.virt.hardware [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 856.343367] env[62952]: DEBUG nova.virt.hardware [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 856.343598] env[62952]: DEBUG nova.virt.hardware [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 856.343835] env[62952]: DEBUG nova.virt.hardware [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 856.344072] env[62952]: DEBUG nova.virt.hardware [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 856.345074] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d93022a-b1d9-4db6-a072-915f844f9171 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.354801] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a630fe-c7ec-4f46-af27-1b58a2891889 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.520572] env[62952]: DEBUG nova.network.neutron [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 856.537170] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263183, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.736031] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.736753] env[62952]: DEBUG nova.compute.manager [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 856.741745] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.335s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.742512] env[62952]: INFO nova.compute.claims [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 856.918223] env[62952]: DEBUG nova.network.neutron [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Updating instance_info_cache with network_info: [{"id": "52b68252-6346-4743-b795-f02ea8effef1", "address": "fa:16:3e:c1:4e:f5", "network": {"id": "0f16fd6d-5891-4df1-b2fb-057e5fc96ed0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-611956326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfecd7aa86044e6bba93d20fb7017e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52b68252-63", "ovs_interfaceid": "52b68252-6346-4743-b795-f02ea8effef1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.940770] env[62952]: DEBUG nova.compute.manager [req-0b030ef6-8a35-4cd5-81a1-5cbfd94bd09f req-fe581cbd-e356-45cd-b2f8-66cc99b724de service nova] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Received event network-vif-plugged-ef78c1ce-2d7f-4e00-87a2-95a93cfc3f37 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 856.941068] env[62952]: DEBUG oslo_concurrency.lockutils [req-0b030ef6-8a35-4cd5-81a1-5cbfd94bd09f req-fe581cbd-e356-45cd-b2f8-66cc99b724de service nova] Acquiring lock "58e703ea-7ccf-4114-a5a2-49533390ed37-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.941315] env[62952]: DEBUG oslo_concurrency.lockutils [req-0b030ef6-8a35-4cd5-81a1-5cbfd94bd09f req-fe581cbd-e356-45cd-b2f8-66cc99b724de service nova] Lock "58e703ea-7ccf-4114-a5a2-49533390ed37-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.941498] env[62952]: DEBUG oslo_concurrency.lockutils [req-0b030ef6-8a35-4cd5-81a1-5cbfd94bd09f req-fe581cbd-e356-45cd-b2f8-66cc99b724de service nova] Lock "58e703ea-7ccf-4114-a5a2-49533390ed37-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.941795] env[62952]: DEBUG nova.compute.manager [req-0b030ef6-8a35-4cd5-81a1-5cbfd94bd09f req-fe581cbd-e356-45cd-b2f8-66cc99b724de service nova] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] No waiting events found dispatching network-vif-plugged-ef78c1ce-2d7f-4e00-87a2-95a93cfc3f37 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 856.941854] env[62952]: WARNING nova.compute.manager [req-0b030ef6-8a35-4cd5-81a1-5cbfd94bd09f req-fe581cbd-e356-45cd-b2f8-66cc99b724de service nova] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Received unexpected event network-vif-plugged-ef78c1ce-2d7f-4e00-87a2-95a93cfc3f37 for instance with vm_state building and task_state spawning. [ 857.028338] env[62952]: DEBUG oslo_vmware.api [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263183, 'name': PowerOnVM_Task, 'duration_secs': 0.535033} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.028624] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 857.028837] env[62952]: INFO nova.compute.manager [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Took 5.86 seconds to spawn the instance on the hypervisor. [ 857.030826] env[62952]: DEBUG nova.compute.manager [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.030826] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2092ba3-1c01-43e6-9733-44cbabe5fffc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.037323] env[62952]: DEBUG nova.network.neutron [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Successfully updated port: ef78c1ce-2d7f-4e00-87a2-95a93cfc3f37 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.247775] env[62952]: DEBUG nova.compute.utils [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 857.251015] env[62952]: DEBUG nova.compute.manager [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 857.251207] env[62952]: DEBUG nova.network.neutron [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 857.290627] env[62952]: DEBUG nova.policy [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '42a664b22bb34b74bd6ed390e40e0b64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfecd7aa86044e6bba93d20fb7017e12', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 857.423128] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Releasing lock "refresh_cache-c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.423462] env[62952]: DEBUG nova.compute.manager [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Instance network_info: |[{"id": "52b68252-6346-4743-b795-f02ea8effef1", "address": "fa:16:3e:c1:4e:f5", "network": {"id": "0f16fd6d-5891-4df1-b2fb-057e5fc96ed0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-611956326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfecd7aa86044e6bba93d20fb7017e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52b68252-63", "ovs_interfaceid": "52b68252-6346-4743-b795-f02ea8effef1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 857.424045] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c1:4e:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49b5df12-d801-4140-8816-2fd401608c7d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '52b68252-6346-4743-b795-f02ea8effef1', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 857.431935] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Creating folder: Project (cfecd7aa86044e6bba93d20fb7017e12). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 857.432234] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f04cff5f-93b4-4467-a92c-e28bb92dc7c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.445170] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Created folder: Project (cfecd7aa86044e6bba93d20fb7017e12) in parent group-v271811. [ 857.445394] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Creating folder: Instances. Parent ref: group-v271851. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 857.445674] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb9935f8-1c16-4772-827f-c87b46689a13 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.457730] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Created folder: Instances in parent group-v271851. [ 857.458031] env[62952]: DEBUG oslo.service.loopingcall [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 857.458269] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 857.458555] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5a322470-0286-425c-b67d-ad429f0707ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.483111] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 857.483111] env[62952]: value = "task-1263186" [ 857.483111] env[62952]: _type = "Task" [ 857.483111] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.492316] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263186, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.540782] env[62952]: DEBUG nova.network.neutron [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Successfully created port: 5805101a-366b-4401-9db9-f69ed60ff18d {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 857.543052] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "refresh_cache-58e703ea-7ccf-4114-a5a2-49533390ed37" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.543199] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquired lock "refresh_cache-58e703ea-7ccf-4114-a5a2-49533390ed37" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.543418] env[62952]: DEBUG nova.network.neutron [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 857.554917] env[62952]: INFO nova.compute.manager [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Took 30.60 seconds to build instance. [ 857.754585] env[62952]: DEBUG nova.compute.manager [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 858.006741] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263186, 'name': CreateVM_Task, 'duration_secs': 0.42186} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.009444] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 858.010529] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.010738] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.011274] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 858.011613] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6e324e1-1535-4b4a-968d-b8e4d0868f88 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.017984] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 858.017984] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52cd9ad7-f7f3-016d-4d83-f4419281ae83" [ 858.017984] env[62952]: _type = "Task" [ 858.017984] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.029834] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52cd9ad7-f7f3-016d-4d83-f4419281ae83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.061084] env[62952]: DEBUG oslo_concurrency.lockutils [None req-220969c7-77ed-40ce-a27a-93b74e6cb967 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Lock "bc369b94-2f40-444e-88d7-069fd33e2646" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.233s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.080541] env[62952]: DEBUG nova.network.neutron [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 858.116941] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eab721d-8d4b-4006-a9d4-29845135d42d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.127824] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80de54a-4d6a-44b0-a5a3-aacd4f58f56d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.167127] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeafca7f-10cc-41e5-bc13-1e9b7613baa3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.177712] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb4407f2-1d23-4969-a525-c2b8d9828613 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.193067] env[62952]: DEBUG nova.compute.provider_tree [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.204845] env[62952]: DEBUG oslo_vmware.rw_handles [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5254cd6e-dfe7-7bde-995d-13fcf171fbc5/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 858.205730] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0037b87-49d2-4a24-bd01-d757c0457896 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.213163] env[62952]: DEBUG oslo_vmware.rw_handles [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5254cd6e-dfe7-7bde-995d-13fcf171fbc5/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 858.213363] env[62952]: ERROR oslo_vmware.rw_handles [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5254cd6e-dfe7-7bde-995d-13fcf171fbc5/disk-0.vmdk due to incomplete transfer. [ 858.213637] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5d61e27f-f608-41ac-bcd8-c6ec10af6ce4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.222174] env[62952]: DEBUG oslo_vmware.rw_handles [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5254cd6e-dfe7-7bde-995d-13fcf171fbc5/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 858.222826] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Uploaded image 8518bf0a-b7da-4699-a0d0-32c8ad4958d8 to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 858.224734] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 858.225474] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d32e169c-f942-48c4-9895-69e0590c3377 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.232876] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 858.232876] env[62952]: value = "task-1263187" [ 858.232876] env[62952]: _type = "Task" [ 858.232876] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.241377] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263187, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.255421] env[62952]: DEBUG nova.network.neutron [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Updating instance_info_cache with network_info: [{"id": "ef78c1ce-2d7f-4e00-87a2-95a93cfc3f37", "address": "fa:16:3e:cb:1e:ba", "network": {"id": "0f16fd6d-5891-4df1-b2fb-057e5fc96ed0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-611956326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfecd7aa86044e6bba93d20fb7017e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef78c1ce-2d", "ovs_interfaceid": "ef78c1ce-2d7f-4e00-87a2-95a93cfc3f37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.421234] env[62952]: INFO nova.compute.manager [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Rebuilding instance [ 858.462061] env[62952]: DEBUG nova.compute.manager [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.463167] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be8741e-b8df-4d4c-bb87-a1f1c2c9af1b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.532041] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52cd9ad7-f7f3-016d-4d83-f4419281ae83, 'name': SearchDatastore_Task, 'duration_secs': 0.015318} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.532562] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.532929] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 858.533328] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.533573] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.533862] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 858.534227] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a10a0172-e790-414d-8124-089af7ad7b39 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.546582] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 858.546855] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 858.548304] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51d3e2c5-b24b-4905-a596-fdb3105ff14d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.554638] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 858.554638] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5254ad9f-7589-1b5f-00b6-5593c9b8e67f" [ 858.554638] env[62952]: _type = "Task" [ 858.554638] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.565452] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5254ad9f-7589-1b5f-00b6-5593c9b8e67f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.565803] env[62952]: DEBUG nova.compute.manager [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 858.698341] env[62952]: DEBUG nova.scheduler.client.report [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 858.747749] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263187, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.758053] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Releasing lock "refresh_cache-58e703ea-7ccf-4114-a5a2-49533390ed37" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.758484] env[62952]: DEBUG nova.compute.manager [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Instance network_info: |[{"id": "ef78c1ce-2d7f-4e00-87a2-95a93cfc3f37", "address": "fa:16:3e:cb:1e:ba", "network": {"id": "0f16fd6d-5891-4df1-b2fb-057e5fc96ed0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-611956326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfecd7aa86044e6bba93d20fb7017e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef78c1ce-2d", "ovs_interfaceid": "ef78c1ce-2d7f-4e00-87a2-95a93cfc3f37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 858.758955] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:1e:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49b5df12-d801-4140-8816-2fd401608c7d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef78c1ce-2d7f-4e00-87a2-95a93cfc3f37', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 858.767218] env[62952]: DEBUG oslo.service.loopingcall [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.768405] env[62952]: DEBUG nova.compute.manager [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 858.770470] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 858.770922] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d9ea8d7-06a6-4b5d-86a5-941a00539bf6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.795983] env[62952]: DEBUG nova.virt.hardware [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.796175] env[62952]: DEBUG nova.virt.hardware [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.796288] env[62952]: DEBUG nova.virt.hardware [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.796443] env[62952]: DEBUG nova.virt.hardware [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.796575] env[62952]: DEBUG nova.virt.hardware [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.796779] env[62952]: DEBUG nova.virt.hardware [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.796942] env[62952]: DEBUG nova.virt.hardware [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.797129] env[62952]: DEBUG nova.virt.hardware [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.797442] env[62952]: DEBUG nova.virt.hardware [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.797530] env[62952]: DEBUG nova.virt.hardware [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.797649] env[62952]: DEBUG nova.virt.hardware [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.798598] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3552dc-6ad9-427f-b7f6-c57bf093ef6f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.802736] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.802736] env[62952]: value = "task-1263188" [ 858.802736] env[62952]: _type = "Task" [ 858.802736] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.814974] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d13ff9e-1930-4ca5-b70e-f18dc1617897 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.819108] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263188, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.973928] env[62952]: DEBUG nova.compute.manager [req-4ddfe4ad-a804-4c5a-b955-1f8f6e01e61d req-0b2729dc-7bad-4b6a-b58c-44355986b947 service nova] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Received event network-changed-ef78c1ce-2d7f-4e00-87a2-95a93cfc3f37 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 858.974156] env[62952]: DEBUG nova.compute.manager [req-4ddfe4ad-a804-4c5a-b955-1f8f6e01e61d req-0b2729dc-7bad-4b6a-b58c-44355986b947 service nova] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Refreshing instance network info cache due to event network-changed-ef78c1ce-2d7f-4e00-87a2-95a93cfc3f37. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 858.974431] env[62952]: DEBUG oslo_concurrency.lockutils [req-4ddfe4ad-a804-4c5a-b955-1f8f6e01e61d req-0b2729dc-7bad-4b6a-b58c-44355986b947 service nova] Acquiring lock "refresh_cache-58e703ea-7ccf-4114-a5a2-49533390ed37" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.974662] env[62952]: DEBUG oslo_concurrency.lockutils [req-4ddfe4ad-a804-4c5a-b955-1f8f6e01e61d req-0b2729dc-7bad-4b6a-b58c-44355986b947 service nova] Acquired lock "refresh_cache-58e703ea-7ccf-4114-a5a2-49533390ed37" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.974755] env[62952]: DEBUG nova.network.neutron [req-4ddfe4ad-a804-4c5a-b955-1f8f6e01e61d req-0b2729dc-7bad-4b6a-b58c-44355986b947 service nova] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Refreshing network info cache for port ef78c1ce-2d7f-4e00-87a2-95a93cfc3f37 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 858.977927] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 858.978608] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a165b9fd-2d1f-4967-ba62-a108c7e94f99 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.005880] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 859.005880] env[62952]: value = "task-1263189" [ 859.005880] env[62952]: _type = "Task" [ 859.005880] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.015645] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263189, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.065182] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5254ad9f-7589-1b5f-00b6-5593c9b8e67f, 'name': SearchDatastore_Task, 'duration_secs': 0.043688} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.066222] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc801f2e-0212-42f2-aae4-e3e2eec63875 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.076689] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 859.076689] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525ccb0c-44e2-6439-a7dc-59afe78be9d7" [ 859.076689] env[62952]: _type = "Task" [ 859.076689] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.090931] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525ccb0c-44e2-6439-a7dc-59afe78be9d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.099017] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.160193] env[62952]: DEBUG nova.network.neutron [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Successfully updated port: 5805101a-366b-4401-9db9-f69ed60ff18d {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 859.203115] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.203542] env[62952]: DEBUG nova.compute.manager [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 859.209109] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.789s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.209109] env[62952]: DEBUG nova.objects.instance [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Lazy-loading 'resources' on Instance uuid 8ac98d96-139a-4a03-94f3-178fafa88503 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 859.252738] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263187, 'name': Destroy_Task, 'duration_secs': 0.998825} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.254226] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Destroyed the VM [ 859.254697] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 859.255456] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-7d7f7329-5c96-4898-8385-49c4e5a94f6d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.265852] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 859.265852] env[62952]: value = "task-1263190" [ 859.265852] env[62952]: _type = "Task" [ 859.265852] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.277115] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263190, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.316051] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263188, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.521263] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263189, 'name': PowerOffVM_Task, 'duration_secs': 0.324135} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.521263] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 859.521263] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 859.521263] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d33c5181-afba-4fb4-ab50-026b714b6dce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.529220] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 859.529496] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-72ee2a67-b92b-4c90-b91b-c3e06040b7f3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.557399] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 859.557810] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 859.558048] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Deleting the datastore file [datastore2] bc369b94-2f40-444e-88d7-069fd33e2646 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 859.558335] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4f864da8-966e-46a9-bcd4-3a9874594219 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.565984] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 859.565984] env[62952]: value = "task-1263192" [ 859.565984] env[62952]: _type = "Task" [ 859.565984] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.578319] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263192, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.588290] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525ccb0c-44e2-6439-a7dc-59afe78be9d7, 'name': SearchDatastore_Task, 'duration_secs': 0.027787} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.588599] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.588870] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb/c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 859.589184] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-410fd0e7-cad2-44fd-bd10-214a3f9954aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.598293] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 859.598293] env[62952]: value = "task-1263193" [ 859.598293] env[62952]: _type = "Task" [ 859.598293] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.606950] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263193, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.663552] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "refresh_cache-f906dc47-12cb-46f5-ae5c-53cfcc8765a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.663723] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquired lock "refresh_cache-f906dc47-12cb-46f5-ae5c-53cfcc8765a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.663885] env[62952]: DEBUG nova.network.neutron [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 859.712060] env[62952]: DEBUG nova.compute.utils [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 859.712909] env[62952]: DEBUG nova.compute.manager [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 859.713462] env[62952]: DEBUG nova.network.neutron [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 859.725449] env[62952]: DEBUG nova.network.neutron [req-4ddfe4ad-a804-4c5a-b955-1f8f6e01e61d req-0b2729dc-7bad-4b6a-b58c-44355986b947 service nova] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Updated VIF entry in instance network info cache for port ef78c1ce-2d7f-4e00-87a2-95a93cfc3f37. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 859.729883] env[62952]: DEBUG nova.network.neutron [req-4ddfe4ad-a804-4c5a-b955-1f8f6e01e61d req-0b2729dc-7bad-4b6a-b58c-44355986b947 service nova] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Updating instance_info_cache with network_info: [{"id": "ef78c1ce-2d7f-4e00-87a2-95a93cfc3f37", "address": "fa:16:3e:cb:1e:ba", "network": {"id": "0f16fd6d-5891-4df1-b2fb-057e5fc96ed0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-611956326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfecd7aa86044e6bba93d20fb7017e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef78c1ce-2d", "ovs_interfaceid": "ef78c1ce-2d7f-4e00-87a2-95a93cfc3f37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.776515] env[62952]: DEBUG nova.policy [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6967f8ebdc944dbfb2d19b5d7acdd3e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91fde3ceb94649f698459dfc174bf835', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 859.784384] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263190, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.816617] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263188, 'name': CreateVM_Task, 'duration_secs': 0.955738} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.816819] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 859.817584] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.817784] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.818152] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.818418] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2e4469d-3f76-47ae-a550-e618d9fa5bd1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.826724] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 859.826724] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]527f18a2-7b4e-4859-7de2-1c188c522609" [ 859.826724] env[62952]: _type = "Task" [ 859.826724] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.840707] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]527f18a2-7b4e-4859-7de2-1c188c522609, 'name': SearchDatastore_Task, 'duration_secs': 0.010332} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.845075] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.845335] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 859.845600] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.845766] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.845969] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.849578] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-75dac132-0dc8-4d64-94a3-bce92e80226e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.851797] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "7a73cb49-3e93-4fcc-b613-3808a8d5c64a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.852068] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "7a73cb49-3e93-4fcc-b613-3808a8d5c64a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.871062] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.871418] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 859.872315] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4994fd5a-cc3e-49db-866b-de7ac4a70c54 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.880210] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 859.880210] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d988ef-75b0-c6f1-827f-89b06028afe3" [ 859.880210] env[62952]: _type = "Task" [ 859.880210] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.895054] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d988ef-75b0-c6f1-827f-89b06028afe3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.077588] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263192, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151163} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.078138] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 860.078138] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 860.078392] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 860.112666] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263193, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.158207] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-550823f8-fc4c-45c8-849d-65434083fd8b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.162357] env[62952]: DEBUG nova.network.neutron [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Successfully created port: b5b040c9-55a9-4929-b720-a1988f7492c9 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 860.170426] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd0ecc6-b228-42d7-8506-f41c4b77591d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.201699] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a9f261b-11f6-4e55-bee0-9bc5b1ccb0f8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.209616] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57dce0f-9b2d-49de-a6d4-6a28bf128b86 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.214106] env[62952]: DEBUG nova.network.neutron [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 860.225777] env[62952]: DEBUG nova.compute.manager [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 860.228630] env[62952]: DEBUG nova.compute.provider_tree [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.234374] env[62952]: DEBUG oslo_concurrency.lockutils [req-4ddfe4ad-a804-4c5a-b955-1f8f6e01e61d req-0b2729dc-7bad-4b6a-b58c-44355986b947 service nova] Releasing lock "refresh_cache-58e703ea-7ccf-4114-a5a2-49533390ed37" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.278812] env[62952]: DEBUG oslo_vmware.api [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263190, 'name': RemoveSnapshot_Task, 'duration_secs': 0.823456} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.279177] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 860.279347] env[62952]: INFO nova.compute.manager [None req-22f668a0-2f94-485f-b54e-d2150f07b1ac tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Took 15.49 seconds to snapshot the instance on the hypervisor. [ 860.390653] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d988ef-75b0-c6f1-827f-89b06028afe3, 'name': SearchDatastore_Task, 'duration_secs': 0.057019} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.393568] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc9c7e93-d6e1-4e37-a6f9-6bdee627aebb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.401030] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 860.401030] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b20f76-9786-b970-327d-84cf6f6873e9" [ 860.401030] env[62952]: _type = "Task" [ 860.401030] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.408822] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b20f76-9786-b970-327d-84cf6f6873e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.461034] env[62952]: DEBUG nova.network.neutron [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Updating instance_info_cache with network_info: [{"id": "5805101a-366b-4401-9db9-f69ed60ff18d", "address": "fa:16:3e:6e:dd:5d", "network": {"id": "0f16fd6d-5891-4df1-b2fb-057e5fc96ed0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-611956326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfecd7aa86044e6bba93d20fb7017e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5805101a-36", "ovs_interfaceid": "5805101a-366b-4401-9db9-f69ed60ff18d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.607781] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263193, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.574799} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.608103] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb/c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 860.608300] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 860.608548] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2c9664c5-b34b-4439-a36a-14cc65bbc540 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.616207] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 860.616207] env[62952]: value = "task-1263194" [ 860.616207] env[62952]: _type = "Task" [ 860.616207] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.625722] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263194, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.736411] env[62952]: DEBUG nova.scheduler.client.report [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.910259] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b20f76-9786-b970-327d-84cf6f6873e9, 'name': SearchDatastore_Task, 'duration_secs': 0.010758} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.910531] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.910792] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 58e703ea-7ccf-4114-a5a2-49533390ed37/58e703ea-7ccf-4114-a5a2-49533390ed37.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 860.911061] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-869699cf-1c00-4b14-851b-66abc9d4bff1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.918255] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 860.918255] env[62952]: value = "task-1263195" [ 860.918255] env[62952]: _type = "Task" [ 860.918255] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.926483] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263195, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.962978] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Releasing lock "refresh_cache-f906dc47-12cb-46f5-ae5c-53cfcc8765a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.963366] env[62952]: DEBUG nova.compute.manager [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Instance network_info: |[{"id": "5805101a-366b-4401-9db9-f69ed60ff18d", "address": "fa:16:3e:6e:dd:5d", "network": {"id": "0f16fd6d-5891-4df1-b2fb-057e5fc96ed0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-611956326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfecd7aa86044e6bba93d20fb7017e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5805101a-36", "ovs_interfaceid": "5805101a-366b-4401-9db9-f69ed60ff18d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 860.963834] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:dd:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49b5df12-d801-4140-8816-2fd401608c7d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5805101a-366b-4401-9db9-f69ed60ff18d', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 860.971886] env[62952]: DEBUG oslo.service.loopingcall [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.972149] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 860.972680] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-24970fb2-676d-4673-b38d-a1c820f68e8a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.995388] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 860.995388] env[62952]: value = "task-1263196" [ 860.995388] env[62952]: _type = "Task" [ 860.995388] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.005247] env[62952]: DEBUG nova.compute.manager [req-31f7cd2d-307a-4728-a261-8e005f3b9ee1 req-3a68ee98-e339-4021-a5dd-fa0774ac6f83 service nova] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Received event network-vif-plugged-5805101a-366b-4401-9db9-f69ed60ff18d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 861.005539] env[62952]: DEBUG oslo_concurrency.lockutils [req-31f7cd2d-307a-4728-a261-8e005f3b9ee1 req-3a68ee98-e339-4021-a5dd-fa0774ac6f83 service nova] Acquiring lock "f906dc47-12cb-46f5-ae5c-53cfcc8765a8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.005797] env[62952]: DEBUG oslo_concurrency.lockutils [req-31f7cd2d-307a-4728-a261-8e005f3b9ee1 req-3a68ee98-e339-4021-a5dd-fa0774ac6f83 service nova] Lock "f906dc47-12cb-46f5-ae5c-53cfcc8765a8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.006071] env[62952]: DEBUG oslo_concurrency.lockutils [req-31f7cd2d-307a-4728-a261-8e005f3b9ee1 req-3a68ee98-e339-4021-a5dd-fa0774ac6f83 service nova] Lock "f906dc47-12cb-46f5-ae5c-53cfcc8765a8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.006346] env[62952]: DEBUG nova.compute.manager [req-31f7cd2d-307a-4728-a261-8e005f3b9ee1 req-3a68ee98-e339-4021-a5dd-fa0774ac6f83 service nova] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] No waiting events found dispatching network-vif-plugged-5805101a-366b-4401-9db9-f69ed60ff18d {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 861.007125] env[62952]: WARNING nova.compute.manager [req-31f7cd2d-307a-4728-a261-8e005f3b9ee1 req-3a68ee98-e339-4021-a5dd-fa0774ac6f83 service nova] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Received unexpected event network-vif-plugged-5805101a-366b-4401-9db9-f69ed60ff18d for instance with vm_state building and task_state spawning. [ 861.007125] env[62952]: DEBUG nova.compute.manager [req-31f7cd2d-307a-4728-a261-8e005f3b9ee1 req-3a68ee98-e339-4021-a5dd-fa0774ac6f83 service nova] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Received event network-changed-5805101a-366b-4401-9db9-f69ed60ff18d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 861.007125] env[62952]: DEBUG nova.compute.manager [req-31f7cd2d-307a-4728-a261-8e005f3b9ee1 req-3a68ee98-e339-4021-a5dd-fa0774ac6f83 service nova] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Refreshing instance network info cache due to event network-changed-5805101a-366b-4401-9db9-f69ed60ff18d. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 861.007304] env[62952]: DEBUG oslo_concurrency.lockutils [req-31f7cd2d-307a-4728-a261-8e005f3b9ee1 req-3a68ee98-e339-4021-a5dd-fa0774ac6f83 service nova] Acquiring lock "refresh_cache-f906dc47-12cb-46f5-ae5c-53cfcc8765a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.007407] env[62952]: DEBUG oslo_concurrency.lockutils [req-31f7cd2d-307a-4728-a261-8e005f3b9ee1 req-3a68ee98-e339-4021-a5dd-fa0774ac6f83 service nova] Acquired lock "refresh_cache-f906dc47-12cb-46f5-ae5c-53cfcc8765a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.007612] env[62952]: DEBUG nova.network.neutron [req-31f7cd2d-307a-4728-a261-8e005f3b9ee1 req-3a68ee98-e339-4021-a5dd-fa0774ac6f83 service nova] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Refreshing network info cache for port 5805101a-366b-4401-9db9-f69ed60ff18d {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 861.012903] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263196, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.116186] env[62952]: DEBUG nova.virt.hardware [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 861.116477] env[62952]: DEBUG nova.virt.hardware [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 861.116662] env[62952]: DEBUG nova.virt.hardware [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 861.116928] env[62952]: DEBUG nova.virt.hardware [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 861.117124] env[62952]: DEBUG nova.virt.hardware [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 861.117287] env[62952]: DEBUG nova.virt.hardware [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 861.117517] env[62952]: DEBUG nova.virt.hardware [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 861.117699] env[62952]: DEBUG nova.virt.hardware [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 861.117873] env[62952]: DEBUG nova.virt.hardware [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 861.118070] env[62952]: DEBUG nova.virt.hardware [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 861.118303] env[62952]: DEBUG nova.virt.hardware [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 861.119253] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46b471e-2c55-4db8-8815-4ab4af853715 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.133031] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99cb2971-ebc9-4980-9e4d-7093e8fd28b4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.136738] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263194, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069943} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.136991] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 861.138044] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202dbaad-5d43-4f62-b04d-db67af5e6796 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.149771] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 861.155330] env[62952]: DEBUG oslo.service.loopingcall [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.155962] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 861.156202] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-982e711c-1e6b-440e-9239-b98fa5ab1f31 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.187702] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb/c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 861.188558] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b618f616-ea0f-4efc-9e16-23b01cc544bd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.204841] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 861.204841] env[62952]: value = "task-1263197" [ 861.204841] env[62952]: _type = "Task" [ 861.204841] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.210716] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 861.210716] env[62952]: value = "task-1263198" [ 861.210716] env[62952]: _type = "Task" [ 861.210716] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.214165] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263197, 'name': CreateVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.224739] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263198, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.243175] env[62952]: DEBUG nova.compute.manager [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 861.245136] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.037s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.248457] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 19.143s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.269157] env[62952]: DEBUG nova.virt.hardware [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 861.269471] env[62952]: DEBUG nova.virt.hardware [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 861.269663] env[62952]: DEBUG nova.virt.hardware [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 861.269876] env[62952]: DEBUG nova.virt.hardware [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 861.270061] env[62952]: DEBUG nova.virt.hardware [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 861.270233] env[62952]: DEBUG nova.virt.hardware [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 861.270466] env[62952]: DEBUG nova.virt.hardware [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 861.270677] env[62952]: DEBUG nova.virt.hardware [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 861.270886] env[62952]: DEBUG nova.virt.hardware [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 861.271196] env[62952]: DEBUG nova.virt.hardware [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 861.271456] env[62952]: DEBUG nova.virt.hardware [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 861.272592] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acbbad74-1929-439f-9477-1c8698d0c114 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.278624] env[62952]: INFO nova.scheduler.client.report [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Deleted allocations for instance 8ac98d96-139a-4a03-94f3-178fafa88503 [ 861.289967] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae972fd-96b8-4764-9840-0715f3f8367b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.429937] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263195, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.507256] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263196, 'name': CreateVM_Task, 'duration_secs': 0.384989} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.507256] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 861.507831] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.508466] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.508932] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 861.511281] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2df3b683-ef02-43e7-9e54-9f15f6f0eb65 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.519125] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 861.519125] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a2512d-3016-829c-0c7b-1997a94de865" [ 861.519125] env[62952]: _type = "Task" [ 861.519125] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.529306] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a2512d-3016-829c-0c7b-1997a94de865, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.716238] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263197, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.726954] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263198, 'name': ReconfigVM_Task, 'duration_secs': 0.372191} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.727341] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Reconfigured VM instance instance-00000036 to attach disk [datastore2] c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb/c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.728345] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0a1b1117-c05f-4c24-9e04-d127bc443ec4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.737100] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 861.737100] env[62952]: value = "task-1263199" [ 861.737100] env[62952]: _type = "Task" [ 861.737100] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.746681] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263199, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.788408] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3d5c68f1-062e-4677-a259-ee56b374cd6c tempest-ImagesNegativeTestJSON-1941628584 tempest-ImagesNegativeTestJSON-1941628584-project-member] Lock "8ac98d96-139a-4a03-94f3-178fafa88503" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.319s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.843503] env[62952]: DEBUG nova.network.neutron [req-31f7cd2d-307a-4728-a261-8e005f3b9ee1 req-3a68ee98-e339-4021-a5dd-fa0774ac6f83 service nova] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Updated VIF entry in instance network info cache for port 5805101a-366b-4401-9db9-f69ed60ff18d. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 861.844053] env[62952]: DEBUG nova.network.neutron [req-31f7cd2d-307a-4728-a261-8e005f3b9ee1 req-3a68ee98-e339-4021-a5dd-fa0774ac6f83 service nova] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Updating instance_info_cache with network_info: [{"id": "5805101a-366b-4401-9db9-f69ed60ff18d", "address": "fa:16:3e:6e:dd:5d", "network": {"id": "0f16fd6d-5891-4df1-b2fb-057e5fc96ed0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-611956326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfecd7aa86044e6bba93d20fb7017e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5805101a-36", "ovs_interfaceid": "5805101a-366b-4401-9db9-f69ed60ff18d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.932223] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263195, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.691737} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.932740] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 58e703ea-7ccf-4114-a5a2-49533390ed37/58e703ea-7ccf-4114-a5a2-49533390ed37.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 861.933017] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 861.933292] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d01876f0-e17e-467a-a01c-b2c820e96071 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.941431] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 861.941431] env[62952]: value = "task-1263200" [ 861.941431] env[62952]: _type = "Task" [ 861.941431] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.952380] env[62952]: DEBUG nova.network.neutron [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Successfully updated port: b5b040c9-55a9-4929-b720-a1988f7492c9 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 861.959753] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263200, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.034666] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a2512d-3016-829c-0c7b-1997a94de865, 'name': SearchDatastore_Task, 'duration_secs': 0.052193} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.036076] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.036076] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 862.036076] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.036076] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.036296] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 862.039075] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-63ffecc5-1a83-4b96-b0db-193b714645c1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.050192] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 862.050436] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 862.051564] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ee3e2b6-8e42-41f0-b4ee-94c4a07ec254 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.059468] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 862.059468] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]522487e3-922a-2146-2f7a-f5d9d8e15aa1" [ 862.059468] env[62952]: _type = "Task" [ 862.059468] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.070128] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]522487e3-922a-2146-2f7a-f5d9d8e15aa1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.217779] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263197, 'name': CreateVM_Task, 'duration_secs': 0.528578} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.217980] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 862.218415] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.218702] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.219035] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.219835] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdeed531-3c1d-4e41-9d20-c46f66ac7bdb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.224956] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 862.224956] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c8ae83-1afb-b820-bb65-2710fe7ab2df" [ 862.224956] env[62952]: _type = "Task" [ 862.224956] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.233172] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c8ae83-1afb-b820-bb65-2710fe7ab2df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.248276] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263199, 'name': Rename_Task, 'duration_secs': 0.143538} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.249323] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 862.249323] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0545be39-9b5f-417e-8b51-6d7b0b46d66b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.255983] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 862.255983] env[62952]: value = "task-1263201" [ 862.255983] env[62952]: _type = "Task" [ 862.255983] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.271614] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263201, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.297064] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance f4f8bee4-bf92-4321-afdc-16f7fb17dbae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 862.297234] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 5f1eca6f-4cc4-483e-b72a-b479378277fe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 862.297359] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 9cc7bc2f-ff22-43cf-be79-ff8c6426115b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 862.297478] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 219cfb14-ca73-45d3-86e8-ed4642d7d480 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 862.297592] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance bc369b94-2f40-444e-88d7-069fd33e2646 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 862.297705] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 862.297816] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 58e703ea-7ccf-4114-a5a2-49533390ed37 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 862.297933] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance f906dc47-12cb-46f5-ae5c-53cfcc8765a8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 862.298050] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 0533ee81-b69d-4f4e-9832-78c5ab4512b1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 862.346364] env[62952]: DEBUG oslo_concurrency.lockutils [req-31f7cd2d-307a-4728-a261-8e005f3b9ee1 req-3a68ee98-e339-4021-a5dd-fa0774ac6f83 service nova] Releasing lock "refresh_cache-f906dc47-12cb-46f5-ae5c-53cfcc8765a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.451955] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263200, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079367} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.452248] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.453093] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a9247c-a58a-4516-8638-cca2265c70e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.455824] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.455956] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.456142] env[62952]: DEBUG nova.network.neutron [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 862.483714] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 58e703ea-7ccf-4114-a5a2-49533390ed37/58e703ea-7ccf-4114-a5a2-49533390ed37.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.484608] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8b9f5a6-5554-4c7b-b577-509fe3e9c59d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.507293] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 862.507293] env[62952]: value = "task-1263202" [ 862.507293] env[62952]: _type = "Task" [ 862.507293] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.525158] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263202, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.573273] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]522487e3-922a-2146-2f7a-f5d9d8e15aa1, 'name': SearchDatastore_Task, 'duration_secs': 0.009344} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.574190] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31b75125-6d12-415f-ac5b-d920183967ff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.581486] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 862.581486] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ab9cde-ff58-3579-b410-22bd3401f5c5" [ 862.581486] env[62952]: _type = "Task" [ 862.581486] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.597205] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ab9cde-ff58-3579-b410-22bd3401f5c5, 'name': SearchDatastore_Task, 'duration_secs': 0.009914} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.597465] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.597716] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] f906dc47-12cb-46f5-ae5c-53cfcc8765a8/f906dc47-12cb-46f5-ae5c-53cfcc8765a8.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 862.598020] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bba599a7-7541-4281-917d-d7f5bcdd3422 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.607459] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 862.607459] env[62952]: value = "task-1263203" [ 862.607459] env[62952]: _type = "Task" [ 862.607459] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.616774] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263203, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.737313] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c8ae83-1afb-b820-bb65-2710fe7ab2df, 'name': SearchDatastore_Task, 'duration_secs': 0.009409} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.737775] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.738114] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 862.738435] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.738624] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.738818] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 862.739143] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d068be4-490f-4137-95ec-884509bc531e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.749295] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 862.749546] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 862.750588] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5cfc46d-de12-4d33-9319-5bbfda1c1120 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.760262] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 862.760262] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5201de78-71fa-4c13-0ba8-7ad7f636ec4c" [ 862.760262] env[62952]: _type = "Task" [ 862.760262] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.780046] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5201de78-71fa-4c13-0ba8-7ad7f636ec4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.785847] env[62952]: DEBUG oslo_vmware.api [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263201, 'name': PowerOnVM_Task, 'duration_secs': 0.456812} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.786419] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 862.786831] env[62952]: INFO nova.compute.manager [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Took 9.12 seconds to spawn the instance on the hypervisor. [ 862.787848] env[62952]: DEBUG nova.compute.manager [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 862.788907] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2665fa5f-b3a4-4b98-8bab-a59c7ce9ca7e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.804393] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 67b729be-fb66-4ed1-bbea-e62216d460d5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 863.016643] env[62952]: DEBUG nova.network.neutron [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 863.024333] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263202, 'name': ReconfigVM_Task, 'duration_secs': 0.290256} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.024837] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 58e703ea-7ccf-4114-a5a2-49533390ed37/58e703ea-7ccf-4114-a5a2-49533390ed37.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 863.025853] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-698ad747-5b65-4067-9e57-1f5ce2dc89bc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.036448] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 863.036448] env[62952]: value = "task-1263204" [ 863.036448] env[62952]: _type = "Task" [ 863.036448] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.042746] env[62952]: DEBUG nova.compute.manager [req-8484b2b6-fd3e-4445-8f20-f365a53413e2 req-aeabb036-8874-4b7c-85ea-55b5ef46c108 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Received event network-vif-plugged-b5b040c9-55a9-4929-b720-a1988f7492c9 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 863.043103] env[62952]: DEBUG oslo_concurrency.lockutils [req-8484b2b6-fd3e-4445-8f20-f365a53413e2 req-aeabb036-8874-4b7c-85ea-55b5ef46c108 service nova] Acquiring lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.043409] env[62952]: DEBUG oslo_concurrency.lockutils [req-8484b2b6-fd3e-4445-8f20-f365a53413e2 req-aeabb036-8874-4b7c-85ea-55b5ef46c108 service nova] Lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.043599] env[62952]: DEBUG oslo_concurrency.lockutils [req-8484b2b6-fd3e-4445-8f20-f365a53413e2 req-aeabb036-8874-4b7c-85ea-55b5ef46c108 service nova] Lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.043833] env[62952]: DEBUG nova.compute.manager [req-8484b2b6-fd3e-4445-8f20-f365a53413e2 req-aeabb036-8874-4b7c-85ea-55b5ef46c108 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] No waiting events found dispatching network-vif-plugged-b5b040c9-55a9-4929-b720-a1988f7492c9 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 863.044070] env[62952]: WARNING nova.compute.manager [req-8484b2b6-fd3e-4445-8f20-f365a53413e2 req-aeabb036-8874-4b7c-85ea-55b5ef46c108 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Received unexpected event network-vif-plugged-b5b040c9-55a9-4929-b720-a1988f7492c9 for instance with vm_state building and task_state spawning. [ 863.044381] env[62952]: DEBUG nova.compute.manager [req-8484b2b6-fd3e-4445-8f20-f365a53413e2 req-aeabb036-8874-4b7c-85ea-55b5ef46c108 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Received event network-changed-b5b040c9-55a9-4929-b720-a1988f7492c9 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 863.044586] env[62952]: DEBUG nova.compute.manager [req-8484b2b6-fd3e-4445-8f20-f365a53413e2 req-aeabb036-8874-4b7c-85ea-55b5ef46c108 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Refreshing instance network info cache due to event network-changed-b5b040c9-55a9-4929-b720-a1988f7492c9. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 863.044839] env[62952]: DEBUG oslo_concurrency.lockutils [req-8484b2b6-fd3e-4445-8f20-f365a53413e2 req-aeabb036-8874-4b7c-85ea-55b5ef46c108 service nova] Acquiring lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.055728] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263204, 'name': Rename_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.119067] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263203, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474631} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.119342] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] f906dc47-12cb-46f5-ae5c-53cfcc8765a8/f906dc47-12cb-46f5-ae5c-53cfcc8765a8.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 863.119586] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 863.119853] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-91974f9d-ccb6-42f3-bc57-5d39d739b56a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.130891] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 863.130891] env[62952]: value = "task-1263205" [ 863.130891] env[62952]: _type = "Task" [ 863.130891] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.142818] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263205, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.255146] env[62952]: DEBUG nova.network.neutron [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Updating instance_info_cache with network_info: [{"id": "b5b040c9-55a9-4929-b720-a1988f7492c9", "address": "fa:16:3e:0b:89:af", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5b040c9-55", "ovs_interfaceid": "b5b040c9-55a9-4929-b720-a1988f7492c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.277134] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5201de78-71fa-4c13-0ba8-7ad7f636ec4c, 'name': SearchDatastore_Task, 'duration_secs': 0.046443} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.278166] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1650d9c-fcab-4132-ad40-86b614894adb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.284134] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 863.284134] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]527e49bb-cf3f-a8ca-5232-b537910739f7" [ 863.284134] env[62952]: _type = "Task" [ 863.284134] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.294113] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]527e49bb-cf3f-a8ca-5232-b537910739f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.313243] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 2b11ad9b-63d8-40da-8f68-3d8260c63bc7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 863.318741] env[62952]: INFO nova.compute.manager [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Took 27.60 seconds to build instance. [ 863.548462] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263204, 'name': Rename_Task, 'duration_secs': 0.153578} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.548604] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 863.549379] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a7957c5-e28a-43e1-9322-94c55609e1c0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.555975] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 863.555975] env[62952]: value = "task-1263206" [ 863.555975] env[62952]: _type = "Task" [ 863.555975] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.564838] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263206, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.641143] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263205, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.12068} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.641425] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 863.642220] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb0392eb-e4a1-4647-a815-e90624c15f2a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.664791] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] f906dc47-12cb-46f5-ae5c-53cfcc8765a8/f906dc47-12cb-46f5-ae5c-53cfcc8765a8.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.665115] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f96a58be-d5b8-4190-9ae1-1ee376764a3d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.687366] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 863.687366] env[62952]: value = "task-1263207" [ 863.687366] env[62952]: _type = "Task" [ 863.687366] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.695837] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263207, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.757721] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.758255] env[62952]: DEBUG nova.compute.manager [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Instance network_info: |[{"id": "b5b040c9-55a9-4929-b720-a1988f7492c9", "address": "fa:16:3e:0b:89:af", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5b040c9-55", "ovs_interfaceid": "b5b040c9-55a9-4929-b720-a1988f7492c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 863.758663] env[62952]: DEBUG oslo_concurrency.lockutils [req-8484b2b6-fd3e-4445-8f20-f365a53413e2 req-aeabb036-8874-4b7c-85ea-55b5ef46c108 service nova] Acquired lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.758915] env[62952]: DEBUG nova.network.neutron [req-8484b2b6-fd3e-4445-8f20-f365a53413e2 req-aeabb036-8874-4b7c-85ea-55b5ef46c108 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Refreshing network info cache for port b5b040c9-55a9-4929-b720-a1988f7492c9 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 863.760550] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:89:af', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '838c9497-35dd-415e-96c7-8dc21b0cd4b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b5b040c9-55a9-4929-b720-a1988f7492c9', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 863.769956] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Creating folder: Project (91fde3ceb94649f698459dfc174bf835). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 863.773721] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-307deb92-cc30-4dc1-8e1c-296d009deec6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.790854] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Created folder: Project (91fde3ceb94649f698459dfc174bf835) in parent group-v271811. [ 863.791885] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Creating folder: Instances. Parent ref: group-v271857. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 863.793226] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b2df4c48-a661-4bdb-a9e8-ab644fec78ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.798785] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]527e49bb-cf3f-a8ca-5232-b537910739f7, 'name': SearchDatastore_Task, 'duration_secs': 0.009595} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.799775] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.800200] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] bc369b94-2f40-444e-88d7-069fd33e2646/bc369b94-2f40-444e-88d7-069fd33e2646.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 863.800514] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ca2f8987-5eba-4ebb-be8a-537192763ec1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.807826] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 863.807826] env[62952]: value = "task-1263210" [ 863.807826] env[62952]: _type = "Task" [ 863.807826] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.809594] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Created folder: Instances in parent group-v271857. [ 863.809886] env[62952]: DEBUG oslo.service.loopingcall [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 863.813337] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 863.817022] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-169e9a5c-7d68-41f6-9e46-0b94b8f177d1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.830715] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 6e289c20-8dd2-4680-b816-9bbe82fe7aec has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 863.832455] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bffadac8-13c8-48c3-8106-8b41870d73bf tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.884s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.837313] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263210, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.843691] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 863.843691] env[62952]: value = "task-1263211" [ 863.843691] env[62952]: _type = "Task" [ 863.843691] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.855134] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263211, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.031234] env[62952]: DEBUG nova.network.neutron [req-8484b2b6-fd3e-4445-8f20-f365a53413e2 req-aeabb036-8874-4b7c-85ea-55b5ef46c108 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Updated VIF entry in instance network info cache for port b5b040c9-55a9-4929-b720-a1988f7492c9. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 864.031775] env[62952]: DEBUG nova.network.neutron [req-8484b2b6-fd3e-4445-8f20-f365a53413e2 req-aeabb036-8874-4b7c-85ea-55b5ef46c108 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Updating instance_info_cache with network_info: [{"id": "b5b040c9-55a9-4929-b720-a1988f7492c9", "address": "fa:16:3e:0b:89:af", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5b040c9-55", "ovs_interfaceid": "b5b040c9-55a9-4929-b720-a1988f7492c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.068240] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263206, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.198744] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263207, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.319603] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263210, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506554} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.319939] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] bc369b94-2f40-444e-88d7-069fd33e2646/bc369b94-2f40-444e-88d7-069fd33e2646.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 864.320224] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 864.320506] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d5cd58cf-4e2d-43a8-bf5f-dd505bfcd5ee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.327619] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 864.327619] env[62952]: value = "task-1263212" [ 864.327619] env[62952]: _type = "Task" [ 864.327619] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.335749] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263212, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.338311] env[62952]: DEBUG nova.compute.manager [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 864.341959] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance dd700ee9-0957-4210-be50-e4c8ac3ca456 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 864.354625] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263211, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.535429] env[62952]: DEBUG oslo_concurrency.lockutils [req-8484b2b6-fd3e-4445-8f20-f365a53413e2 req-aeabb036-8874-4b7c-85ea-55b5ef46c108 service nova] Releasing lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.569758] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263206, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.699203] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263207, 'name': ReconfigVM_Task, 'duration_secs': 0.595133} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.699541] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Reconfigured VM instance instance-00000038 to attach disk [datastore2] f906dc47-12cb-46f5-ae5c-53cfcc8765a8/f906dc47-12cb-46f5-ae5c-53cfcc8765a8.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.700216] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-84e06680-b889-4265-808d-652436f96c2d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.706712] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 864.706712] env[62952]: value = "task-1263213" [ 864.706712] env[62952]: _type = "Task" [ 864.706712] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.714712] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263213, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.836784] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263212, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067303} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.837329] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 864.838173] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbb8212d-0b00-4491-8943-cc3b720131d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.850322] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 3416eb38-b961-4d48-8b37-8b7b92375dc9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 864.863106] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] bc369b94-2f40-444e-88d7-069fd33e2646/bc369b94-2f40-444e-88d7-069fd33e2646.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 864.866301] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 49947bde-069b-459c-ae19-ae5f090f535b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 864.870248] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56e15210-7d68-4751-bc06-e038cc824134 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.885808] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance ba022bea-e357-4541-a573-d1da2eb49b46 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 864.894903] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263211, 'name': CreateVM_Task, 'duration_secs': 0.59854} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.897382] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 864.897382] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 864.897382] env[62952]: value = "task-1263214" [ 864.897382] env[62952]: _type = "Task" [ 864.897382] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.897382] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.897601] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.898469] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 864.898992] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc5c7780-f640-4f23-b839-0cc8e27b4f2e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.905986] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.907719] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 864.907719] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520a601d-48fe-885b-5ff4-c62079619b2f" [ 864.907719] env[62952]: _type = "Task" [ 864.907719] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.915021] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263214, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.922050] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520a601d-48fe-885b-5ff4-c62079619b2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.069879] env[62952]: DEBUG oslo_vmware.api [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263206, 'name': PowerOnVM_Task, 'duration_secs': 1.057757} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.070202] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 865.070489] env[62952]: INFO nova.compute.manager [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Took 8.75 seconds to spawn the instance on the hypervisor. [ 865.070734] env[62952]: DEBUG nova.compute.manager [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 865.071548] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63aa0bff-56f6-41b0-a2be-94ecea6249ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.216772] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263213, 'name': Rename_Task, 'duration_secs': 0.299651} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.217092] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 865.217356] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fcb41e19-70b4-454b-b372-ab850eb69255 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.223488] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 865.223488] env[62952]: value = "task-1263215" [ 865.223488] env[62952]: _type = "Task" [ 865.223488] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.231702] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263215, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.389044] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 14f8dbbe-2c4a-4948-bf15-106d5b1b0677 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 865.413829] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263214, 'name': ReconfigVM_Task, 'duration_secs': 0.257613} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.417047] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Reconfigured VM instance instance-00000035 to attach disk [datastore2] bc369b94-2f40-444e-88d7-069fd33e2646/bc369b94-2f40-444e-88d7-069fd33e2646.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.417932] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e11b90bf-d7a9-479b-8be6-096e890a2dfa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.428349] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520a601d-48fe-885b-5ff4-c62079619b2f, 'name': SearchDatastore_Task, 'duration_secs': 0.014109} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.431097] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.431377] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.431676] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.431829] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.432061] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 865.432404] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 865.432404] env[62952]: value = "task-1263216" [ 865.432404] env[62952]: _type = "Task" [ 865.432404] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.432598] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f646bb43-bd2f-4425-ae5d-58933c2b6f2f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.448834] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263216, 'name': Rename_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.450298] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 865.450661] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 865.451271] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a202879e-0b94-4f18-a64f-c26336d9ff2a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.458325] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 865.458325] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52bdac1d-818a-289f-6898-afb49c29f771" [ 865.458325] env[62952]: _type = "Task" [ 865.458325] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.467207] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52bdac1d-818a-289f-6898-afb49c29f771, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.589547] env[62952]: INFO nova.compute.manager [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Took 26.53 seconds to build instance. [ 865.734739] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263215, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.893410] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 865.944318] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263216, 'name': Rename_Task, 'duration_secs': 0.126932} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.944653] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 865.944995] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5fdce278-8698-4017-8298-abd56d43b9bc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.952185] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 865.952185] env[62952]: value = "task-1263217" [ 865.952185] env[62952]: _type = "Task" [ 865.952185] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.960441] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263217, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.968641] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52bdac1d-818a-289f-6898-afb49c29f771, 'name': SearchDatastore_Task, 'duration_secs': 0.00959} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.969407] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2f920a7-61e5-4db8-865d-8e4b503cd143 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.974619] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 865.974619] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52aaceaf-3706-34e9-fb5c-c39ab1df61bd" [ 865.974619] env[62952]: _type = "Task" [ 865.974619] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.984851] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52aaceaf-3706-34e9-fb5c-c39ab1df61bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.091563] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2951786f-2fa2-4525-b6f4-6c7f5687dee9 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "58e703ea-7ccf-4114-a5a2-49533390ed37" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.928s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.234909] env[62952]: DEBUG oslo_vmware.api [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263215, 'name': PowerOnVM_Task, 'duration_secs': 0.622229} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.234992] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 866.235470] env[62952]: INFO nova.compute.manager [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Took 7.47 seconds to spawn the instance on the hypervisor. [ 866.235470] env[62952]: DEBUG nova.compute.manager [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.236956] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a047844d-4bb1-489d-a237-ba6290f60786 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.397637] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 86ec7b3e-34d6-45e8-87b2-c8550f035ed0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 866.465520] env[62952]: DEBUG oslo_vmware.api [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263217, 'name': PowerOnVM_Task, 'duration_secs': 0.479221} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.466247] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 866.466247] env[62952]: DEBUG nova.compute.manager [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.466959] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76221de2-a4d0-4c20-8356-16662508c07e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.483894] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52aaceaf-3706-34e9-fb5c-c39ab1df61bd, 'name': SearchDatastore_Task, 'duration_secs': 0.014691} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.484192] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.484459] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 0533ee81-b69d-4f4e-9832-78c5ab4512b1/0533ee81-b69d-4f4e-9832-78c5ab4512b1.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 866.484714] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-965a64e6-7b7c-421f-b6ec-cfd1a41b4d5d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.490847] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 866.490847] env[62952]: value = "task-1263218" [ 866.490847] env[62952]: _type = "Task" [ 866.490847] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.500043] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263218, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.595083] env[62952]: DEBUG nova.compute.manager [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 866.755213] env[62952]: INFO nova.compute.manager [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Took 26.84 seconds to build instance. [ 866.906334] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 47b57348-40d1-4735-9f5a-a7288f8f12ee has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 866.996188] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.000970] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263218, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.120237] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.256773] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ac63cadd-8336-4f1f-ab32-fa254a932da2 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "f906dc47-12cb-46f5-ae5c-53cfcc8765a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.488s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.342661] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquiring lock "bc369b94-2f40-444e-88d7-069fd33e2646" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.343063] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Lock "bc369b94-2f40-444e-88d7-069fd33e2646" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.343321] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquiring lock "bc369b94-2f40-444e-88d7-069fd33e2646-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.343555] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Lock "bc369b94-2f40-444e-88d7-069fd33e2646-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.343773] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Lock "bc369b94-2f40-444e-88d7-069fd33e2646-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.351297] env[62952]: INFO nova.compute.manager [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Terminating instance [ 867.355394] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquiring lock "refresh_cache-bc369b94-2f40-444e-88d7-069fd33e2646" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.356266] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquired lock "refresh_cache-bc369b94-2f40-444e-88d7-069fd33e2646" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.359176] env[62952]: DEBUG nova.network.neutron [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 867.410240] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance e5e7b2b5-34cf-4d37-87ff-1b57264a3516 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 867.468989] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Acquiring lock "45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.469245] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Lock "45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.507082] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263218, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.779905} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.507851] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 0533ee81-b69d-4f4e-9832-78c5ab4512b1/0533ee81-b69d-4f4e-9832-78c5ab4512b1.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 867.508143] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 867.508412] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7c1f58cb-b849-4d54-a8cb-93330ec69838 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.515235] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 867.515235] env[62952]: value = "task-1263219" [ 867.515235] env[62952]: _type = "Task" [ 867.515235] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.523869] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263219, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.762836] env[62952]: DEBUG nova.compute.manager [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 867.878362] env[62952]: DEBUG nova.network.neutron [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 867.915722] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 051f9279-533a-4545-b9e2-272bf40956f6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 867.932316] env[62952]: DEBUG nova.network.neutron [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.027169] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263219, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06685} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.027169] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 868.027472] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322162f0-2ffa-4dad-9d35-37c0b28481c9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.050200] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 0533ee81-b69d-4f4e-9832-78c5ab4512b1/0533ee81-b69d-4f4e-9832-78c5ab4512b1.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 868.050682] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-453a16ab-842b-4eab-a6a7-535debe19090 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.071057] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 868.071057] env[62952]: value = "task-1263220" [ 868.071057] env[62952]: _type = "Task" [ 868.071057] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.081330] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263220, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.291994] env[62952]: DEBUG oslo_concurrency.lockutils [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.419340] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 7a73cb49-3e93-4fcc-b613-3808a8d5c64a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 868.420021] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 868.420178] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2304MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 868.434354] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Releasing lock "refresh_cache-bc369b94-2f40-444e-88d7-069fd33e2646" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.434838] env[62952]: DEBUG nova.compute.manager [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 868.434996] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 868.436029] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-420af075-9a3d-4677-918e-168986a77293 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.444225] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 868.446927] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1202d448-0b02-4762-8a0c-8d6f6b161825 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.453314] env[62952]: DEBUG oslo_vmware.api [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 868.453314] env[62952]: value = "task-1263221" [ 868.453314] env[62952]: _type = "Task" [ 868.453314] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.461153] env[62952]: DEBUG oslo_vmware.api [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263221, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.582010] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263220, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.754467] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc0d75f0-fe8c-401a-bf4d-2a435d6be681 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.764979] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5083820-fd73-4577-a900-fe52f101c825 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.795478] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9bf3d27-ec7f-40e7-9287-cb9df70d515f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.803439] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b7b507e-2b0b-44ce-bf5b-4badb2d05771 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.817564] env[62952]: DEBUG nova.compute.provider_tree [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.964306] env[62952]: DEBUG oslo_vmware.api [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263221, 'name': PowerOffVM_Task, 'duration_secs': 0.12875} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.964643] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 868.964769] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 868.965035] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cd9069be-cc9f-4682-9fab-6d424b51e1e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.993030] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 868.993030] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 868.993030] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Deleting the datastore file [datastore2] bc369b94-2f40-444e-88d7-069fd33e2646 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 868.993030] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc23c731-fca1-4ff8-99e7-ceb1821250ee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.999165] env[62952]: DEBUG oslo_vmware.api [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for the task: (returnval){ [ 868.999165] env[62952]: value = "task-1263223" [ 868.999165] env[62952]: _type = "Task" [ 868.999165] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.008105] env[62952]: DEBUG oslo_vmware.api [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263223, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.084954] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263220, 'name': ReconfigVM_Task, 'duration_secs': 0.638761} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.086138] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 0533ee81-b69d-4f4e-9832-78c5ab4512b1/0533ee81-b69d-4f4e-9832-78c5ab4512b1.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 869.086261] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-839fcec8-a3f9-40c1-b6c6-71278a1d70a6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.095601] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 869.095601] env[62952]: value = "task-1263224" [ 869.095601] env[62952]: _type = "Task" [ 869.095601] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.105239] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263224, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.321166] env[62952]: DEBUG nova.scheduler.client.report [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 869.516211] env[62952]: DEBUG oslo_vmware.api [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Task: {'id': task-1263223, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.274273} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.516517] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 869.516685] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 869.516869] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 869.517181] env[62952]: INFO nova.compute.manager [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Took 1.08 seconds to destroy the instance on the hypervisor. [ 869.517519] env[62952]: DEBUG oslo.service.loopingcall [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.517769] env[62952]: DEBUG nova.compute.manager [-] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 869.517909] env[62952]: DEBUG nova.network.neutron [-] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 869.537180] env[62952]: DEBUG nova.network.neutron [-] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 869.606960] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263224, 'name': Rename_Task, 'duration_secs': 0.311717} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.607328] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 869.607596] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-db1fd79f-944b-4e6d-8df5-91559dc8bd31 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.616089] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 869.616089] env[62952]: value = "task-1263225" [ 869.616089] env[62952]: _type = "Task" [ 869.616089] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.628290] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263225, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.826323] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 869.826608] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.578s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.826905] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.492s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.828778] env[62952]: INFO nova.compute.claims [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.832456] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 869.832632] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Cleaning up deleted instances {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 869.996331] env[62952]: DEBUG oslo_concurrency.lockutils [None req-cbaa7363-73e0-48c2-aba0-04b33b281971 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.996664] env[62952]: DEBUG oslo_concurrency.lockutils [None req-cbaa7363-73e0-48c2-aba0-04b33b281971 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.996794] env[62952]: DEBUG nova.compute.manager [None req-cbaa7363-73e0-48c2-aba0-04b33b281971 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.997838] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4f60b8-5148-43bd-a6dd-bd8a7c1c16d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.006112] env[62952]: DEBUG nova.compute.manager [None req-cbaa7363-73e0-48c2-aba0-04b33b281971 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62952) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 870.006737] env[62952]: DEBUG nova.objects.instance [None req-cbaa7363-73e0-48c2-aba0-04b33b281971 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lazy-loading 'flavor' on Instance uuid c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.040164] env[62952]: DEBUG nova.network.neutron [-] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.126789] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263225, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.339592] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] There are 5 instances to clean {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 870.340074] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 8ac98d96-139a-4a03-94f3-178fafa88503] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 870.512680] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbaa7363-73e0-48c2-aba0-04b33b281971 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 870.512680] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e5786386-84f3-4dcc-a82f-1fc9ecda6e21 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.519172] env[62952]: DEBUG oslo_vmware.api [None req-cbaa7363-73e0-48c2-aba0-04b33b281971 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 870.519172] env[62952]: value = "task-1263226" [ 870.519172] env[62952]: _type = "Task" [ 870.519172] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.531359] env[62952]: DEBUG oslo_vmware.api [None req-cbaa7363-73e0-48c2-aba0-04b33b281971 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263226, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.544198] env[62952]: INFO nova.compute.manager [-] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Took 1.02 seconds to deallocate network for instance. [ 870.629199] env[62952]: DEBUG oslo_vmware.api [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263225, 'name': PowerOnVM_Task, 'duration_secs': 0.590342} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.629481] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 870.629688] env[62952]: INFO nova.compute.manager [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Took 9.39 seconds to spawn the instance on the hypervisor. [ 870.629879] env[62952]: DEBUG nova.compute.manager [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 870.630715] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b05514a-1dc5-4c0e-ad73-aef01eed2cc2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.844134] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 472e60f1-9b6c-415e-8a92-19235ad2e547] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 871.028052] env[62952]: DEBUG oslo_vmware.api [None req-cbaa7363-73e0-48c2-aba0-04b33b281971 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263226, 'name': PowerOffVM_Task, 'duration_secs': 0.280919} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.030296] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbaa7363-73e0-48c2-aba0-04b33b281971 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 871.030551] env[62952]: DEBUG nova.compute.manager [None req-cbaa7363-73e0-48c2-aba0-04b33b281971 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 871.031543] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7250d810-d862-48f9-a5c9-e8a4100a0874 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.052461] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.116606] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72dd22e-731b-443d-95dd-ff5914e91986 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.124038] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af94e977-a3ec-46d3-b2a0-0b7f8a56389e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.157828] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26dcd77e-b0c1-4df2-a039-7663520b1677 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.163110] env[62952]: INFO nova.compute.manager [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Took 29.78 seconds to build instance. [ 871.167227] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee7016d-ae85-4f2d-b73a-a17e436f10e4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.181347] env[62952]: DEBUG nova.compute.provider_tree [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.348371] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 80a30ea8-e260-4b92-9099-be374333896f] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 871.544677] env[62952]: DEBUG oslo_concurrency.lockutils [None req-cbaa7363-73e0-48c2-aba0-04b33b281971 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.548s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.664958] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9c8e74a0-55e3-4e53-8e04-35f5a7a62547 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.654s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.685027] env[62952]: DEBUG nova.scheduler.client.report [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.853629] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: f075161d-7447-4601-a06c-514fc4dd2eb9] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 871.908162] env[62952]: DEBUG nova.compute.manager [req-16b4a986-91b5-4b20-8520-3eed4b86f7fa req-dd90fd3c-4c40-44a0-a97a-abfd28be8595 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Received event network-changed-b5b040c9-55a9-4929-b720-a1988f7492c9 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 871.908162] env[62952]: DEBUG nova.compute.manager [req-16b4a986-91b5-4b20-8520-3eed4b86f7fa req-dd90fd3c-4c40-44a0-a97a-abfd28be8595 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Refreshing instance network info cache due to event network-changed-b5b040c9-55a9-4929-b720-a1988f7492c9. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 871.908162] env[62952]: DEBUG oslo_concurrency.lockutils [req-16b4a986-91b5-4b20-8520-3eed4b86f7fa req-dd90fd3c-4c40-44a0-a97a-abfd28be8595 service nova] Acquiring lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.908162] env[62952]: DEBUG oslo_concurrency.lockutils [req-16b4a986-91b5-4b20-8520-3eed4b86f7fa req-dd90fd3c-4c40-44a0-a97a-abfd28be8595 service nova] Acquired lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.908162] env[62952]: DEBUG nova.network.neutron [req-16b4a986-91b5-4b20-8520-3eed4b86f7fa req-dd90fd3c-4c40-44a0-a97a-abfd28be8595 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Refreshing network info cache for port b5b040c9-55a9-4929-b720-a1988f7492c9 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 872.170432] env[62952]: DEBUG nova.compute.manager [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 872.190534] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.191086] env[62952]: DEBUG nova.compute.manager [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 872.194427] env[62952]: DEBUG oslo_concurrency.lockutils [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.033s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.196144] env[62952]: INFO nova.compute.claims [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.357411] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: dc025711-45be-49b6-a765-f7687b94e6a1] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 872.568948] env[62952]: DEBUG nova.objects.instance [None req-e1b769ac-e68d-4097-99e1-b4e5f17c7796 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lazy-loading 'flavor' on Instance uuid c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 872.666389] env[62952]: DEBUG nova.network.neutron [req-16b4a986-91b5-4b20-8520-3eed4b86f7fa req-dd90fd3c-4c40-44a0-a97a-abfd28be8595 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Updated VIF entry in instance network info cache for port b5b040c9-55a9-4929-b720-a1988f7492c9. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 872.666761] env[62952]: DEBUG nova.network.neutron [req-16b4a986-91b5-4b20-8520-3eed4b86f7fa req-dd90fd3c-4c40-44a0-a97a-abfd28be8595 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Updating instance_info_cache with network_info: [{"id": "b5b040c9-55a9-4929-b720-a1988f7492c9", "address": "fa:16:3e:0b:89:af", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5b040c9-55", "ovs_interfaceid": "b5b040c9-55a9-4929-b720-a1988f7492c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.690089] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.702086] env[62952]: DEBUG nova.compute.utils [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.705751] env[62952]: DEBUG nova.compute.manager [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 872.706498] env[62952]: DEBUG nova.network.neutron [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 872.744865] env[62952]: DEBUG nova.policy [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '676394eb52704a99a27c37c95d60427c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4484b1cb78ff4e7baac4b9a5d1e63549', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 872.861046] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 872.861046] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Cleaning up deleted instances with incomplete migration {{(pid=62952) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 872.891369] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.891652] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.016206] env[62952]: DEBUG nova.network.neutron [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Successfully created port: 3423511f-8828-4fca-aec2-79e921a53579 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 873.076369] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e1b769ac-e68d-4097-99e1-b4e5f17c7796 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "refresh_cache-c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.076589] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e1b769ac-e68d-4097-99e1-b4e5f17c7796 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquired lock "refresh_cache-c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.076903] env[62952]: DEBUG nova.network.neutron [None req-e1b769ac-e68d-4097-99e1-b4e5f17c7796 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 873.077247] env[62952]: DEBUG nova.objects.instance [None req-e1b769ac-e68d-4097-99e1-b4e5f17c7796 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lazy-loading 'info_cache' on Instance uuid c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.169043] env[62952]: DEBUG oslo_concurrency.lockutils [req-16b4a986-91b5-4b20-8520-3eed4b86f7fa req-dd90fd3c-4c40-44a0-a97a-abfd28be8595 service nova] Releasing lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.209666] env[62952]: DEBUG nova.compute.manager [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 873.365208] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.522959] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86115971-664e-494c-a4c0-02e6e835f4cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.532170] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-090e5376-ddff-4e17-bb31-838c627c72db {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.563246] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09097abd-c31e-4485-9d7b-361163a06b33 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.571921] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae3d866-1d31-4feb-846a-2b3d7f0b070c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.589131] env[62952]: DEBUG nova.objects.base [None req-e1b769ac-e68d-4097-99e1-b4e5f17c7796 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 873.589131] env[62952]: DEBUG nova.compute.provider_tree [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.094287] env[62952]: DEBUG nova.scheduler.client.report [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.219088] env[62952]: DEBUG nova.compute.manager [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 874.244597] env[62952]: DEBUG nova.virt.hardware [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.244865] env[62952]: DEBUG nova.virt.hardware [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.245044] env[62952]: DEBUG nova.virt.hardware [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.245241] env[62952]: DEBUG nova.virt.hardware [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.245395] env[62952]: DEBUG nova.virt.hardware [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.245549] env[62952]: DEBUG nova.virt.hardware [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.245756] env[62952]: DEBUG nova.virt.hardware [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.245917] env[62952]: DEBUG nova.virt.hardware [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.246096] env[62952]: DEBUG nova.virt.hardware [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.246262] env[62952]: DEBUG nova.virt.hardware [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.246441] env[62952]: DEBUG nova.virt.hardware [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.247298] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d324edd2-ee1e-414f-81ae-5ba6fe5527ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.255758] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e9330b-ab3d-446a-a414-9efaffbc1b26 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.312866] env[62952]: DEBUG nova.network.neutron [None req-e1b769ac-e68d-4097-99e1-b4e5f17c7796 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Updating instance_info_cache with network_info: [{"id": "52b68252-6346-4743-b795-f02ea8effef1", "address": "fa:16:3e:c1:4e:f5", "network": {"id": "0f16fd6d-5891-4df1-b2fb-057e5fc96ed0", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-611956326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfecd7aa86044e6bba93d20fb7017e12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52b68252-63", "ovs_interfaceid": "52b68252-6346-4743-b795-f02ea8effef1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.439122] env[62952]: DEBUG nova.compute.manager [req-695b2358-27df-4bea-b5f3-bb6076d01e99 req-142c10f6-02ca-4214-8e5b-ad0d934528dc service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Received event network-vif-plugged-3423511f-8828-4fca-aec2-79e921a53579 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 874.439122] env[62952]: DEBUG oslo_concurrency.lockutils [req-695b2358-27df-4bea-b5f3-bb6076d01e99 req-142c10f6-02ca-4214-8e5b-ad0d934528dc service nova] Acquiring lock "67b729be-fb66-4ed1-bbea-e62216d460d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.439122] env[62952]: DEBUG oslo_concurrency.lockutils [req-695b2358-27df-4bea-b5f3-bb6076d01e99 req-142c10f6-02ca-4214-8e5b-ad0d934528dc service nova] Lock "67b729be-fb66-4ed1-bbea-e62216d460d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.439122] env[62952]: DEBUG oslo_concurrency.lockutils [req-695b2358-27df-4bea-b5f3-bb6076d01e99 req-142c10f6-02ca-4214-8e5b-ad0d934528dc service nova] Lock "67b729be-fb66-4ed1-bbea-e62216d460d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.439122] env[62952]: DEBUG nova.compute.manager [req-695b2358-27df-4bea-b5f3-bb6076d01e99 req-142c10f6-02ca-4214-8e5b-ad0d934528dc service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] No waiting events found dispatching network-vif-plugged-3423511f-8828-4fca-aec2-79e921a53579 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 874.439447] env[62952]: WARNING nova.compute.manager [req-695b2358-27df-4bea-b5f3-bb6076d01e99 req-142c10f6-02ca-4214-8e5b-ad0d934528dc service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Received unexpected event network-vif-plugged-3423511f-8828-4fca-aec2-79e921a53579 for instance with vm_state building and task_state spawning. [ 874.522184] env[62952]: DEBUG nova.network.neutron [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Successfully updated port: 3423511f-8828-4fca-aec2-79e921a53579 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 874.599832] env[62952]: DEBUG oslo_concurrency.lockutils [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.405s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.601022] env[62952]: DEBUG nova.compute.manager [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 874.603669] env[62952]: DEBUG oslo_concurrency.lockutils [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.321s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.605183] env[62952]: INFO nova.compute.claims [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 874.816027] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e1b769ac-e68d-4097-99e1-b4e5f17c7796 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Releasing lock "refresh_cache-c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.024739] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Acquiring lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.024889] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Acquired lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.025062] env[62952]: DEBUG nova.network.neutron [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 875.105330] env[62952]: DEBUG nova.compute.utils [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 875.106731] env[62952]: DEBUG nova.compute.manager [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 875.106904] env[62952]: DEBUG nova.network.neutron [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 875.148120] env[62952]: DEBUG nova.policy [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c5cab9fec6604a0881d677df17dda195', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c51c89d19ed4111a5c1631f0ad47cba', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 875.319598] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1b769ac-e68d-4097-99e1-b4e5f17c7796 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 875.319919] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-777f8791-23df-4641-81ce-4801b3c66c13 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.328440] env[62952]: DEBUG oslo_vmware.api [None req-e1b769ac-e68d-4097-99e1-b4e5f17c7796 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 875.328440] env[62952]: value = "task-1263227" [ 875.328440] env[62952]: _type = "Task" [ 875.328440] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.336112] env[62952]: DEBUG oslo_vmware.api [None req-e1b769ac-e68d-4097-99e1-b4e5f17c7796 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263227, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.388849] env[62952]: DEBUG nova.network.neutron [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Successfully created port: 484ad815-dff6-400c-93dd-babc8b8b5c9c {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 875.572257] env[62952]: DEBUG nova.network.neutron [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 875.610704] env[62952]: DEBUG nova.compute.manager [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 875.762527] env[62952]: DEBUG nova.network.neutron [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Updating instance_info_cache with network_info: [{"id": "3423511f-8828-4fca-aec2-79e921a53579", "address": "fa:16:3e:3b:bd:29", "network": {"id": "d865ccd7-b8a3-49d7-811f-6d4466bbbf45", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1549996301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4484b1cb78ff4e7baac4b9a5d1e63549", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a5569c57-a27b-4d4d-a519-784be105114f", "external-id": "nsx-vlan-transportzone-640", "segmentation_id": 640, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3423511f-88", "ovs_interfaceid": "3423511f-8828-4fca-aec2-79e921a53579", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.838251] env[62952]: DEBUG oslo_vmware.api [None req-e1b769ac-e68d-4097-99e1-b4e5f17c7796 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263227, 'name': PowerOnVM_Task, 'duration_secs': 0.382548} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.840642] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1b769ac-e68d-4097-99e1-b4e5f17c7796 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 875.840873] env[62952]: DEBUG nova.compute.manager [None req-e1b769ac-e68d-4097-99e1-b4e5f17c7796 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 875.841969] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8213dcb0-48ff-4207-864d-ffae8f7c7405 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.952441] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8feb614a-5073-40ea-b095-fe62e6ef7a1e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.960469] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9ba3e10-6b6b-45c1-aaf8-4c1e8e42e99e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.990810] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94194708-2763-4545-a263-e01a64d0cff4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.999375] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-588cd4bd-7076-445b-adf0-479b99727c90 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.012145] env[62952]: DEBUG nova.compute.provider_tree [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.266020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Releasing lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.266391] env[62952]: DEBUG nova.compute.manager [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Instance network_info: |[{"id": "3423511f-8828-4fca-aec2-79e921a53579", "address": "fa:16:3e:3b:bd:29", "network": {"id": "d865ccd7-b8a3-49d7-811f-6d4466bbbf45", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1549996301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4484b1cb78ff4e7baac4b9a5d1e63549", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a5569c57-a27b-4d4d-a519-784be105114f", "external-id": "nsx-vlan-transportzone-640", "segmentation_id": 640, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3423511f-88", "ovs_interfaceid": "3423511f-8828-4fca-aec2-79e921a53579", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 876.266826] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3b:bd:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a5569c57-a27b-4d4d-a519-784be105114f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3423511f-8828-4fca-aec2-79e921a53579', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.274780] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Creating folder: Project (4484b1cb78ff4e7baac4b9a5d1e63549). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 876.275074] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-82fe3de7-c7a9-4fa0-b50b-7d25191ae827 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.286510] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Created folder: Project (4484b1cb78ff4e7baac4b9a5d1e63549) in parent group-v271811. [ 876.286699] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Creating folder: Instances. Parent ref: group-v271860. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 876.286921] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-55475d11-51e4-4fab-bbfb-2ffb620e3136 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.296025] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Created folder: Instances in parent group-v271860. [ 876.296025] env[62952]: DEBUG oslo.service.loopingcall [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.296181] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 876.296376] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ff22b886-d22e-4555-b0fd-274d5a7d661c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.313739] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.313739] env[62952]: value = "task-1263230" [ 876.313739] env[62952]: _type = "Task" [ 876.313739] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.321151] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263230, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.465593] env[62952]: DEBUG nova.compute.manager [req-c7b7823b-7007-4092-a461-363e620dfb80 req-853693a0-77ae-412c-b098-ebb1fff28988 service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Received event network-changed-3423511f-8828-4fca-aec2-79e921a53579 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 876.465795] env[62952]: DEBUG nova.compute.manager [req-c7b7823b-7007-4092-a461-363e620dfb80 req-853693a0-77ae-412c-b098-ebb1fff28988 service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Refreshing instance network info cache due to event network-changed-3423511f-8828-4fca-aec2-79e921a53579. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 876.466023] env[62952]: DEBUG oslo_concurrency.lockutils [req-c7b7823b-7007-4092-a461-363e620dfb80 req-853693a0-77ae-412c-b098-ebb1fff28988 service nova] Acquiring lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.466178] env[62952]: DEBUG oslo_concurrency.lockutils [req-c7b7823b-7007-4092-a461-363e620dfb80 req-853693a0-77ae-412c-b098-ebb1fff28988 service nova] Acquired lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.466341] env[62952]: DEBUG nova.network.neutron [req-c7b7823b-7007-4092-a461-363e620dfb80 req-853693a0-77ae-412c-b098-ebb1fff28988 service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Refreshing network info cache for port 3423511f-8828-4fca-aec2-79e921a53579 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 876.515947] env[62952]: DEBUG nova.scheduler.client.report [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.629781] env[62952]: DEBUG nova.compute.manager [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 876.655853] env[62952]: DEBUG nova.virt.hardware [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 876.656175] env[62952]: DEBUG nova.virt.hardware [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 876.656348] env[62952]: DEBUG nova.virt.hardware [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.656551] env[62952]: DEBUG nova.virt.hardware [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 876.656713] env[62952]: DEBUG nova.virt.hardware [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.656867] env[62952]: DEBUG nova.virt.hardware [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 876.657104] env[62952]: DEBUG nova.virt.hardware [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 876.657291] env[62952]: DEBUG nova.virt.hardware [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 876.657496] env[62952]: DEBUG nova.virt.hardware [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 876.657682] env[62952]: DEBUG nova.virt.hardware [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 876.657883] env[62952]: DEBUG nova.virt.hardware [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.658823] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dfdc709-c440-4b31-9fb9-1394599f900b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.668857] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-872d849c-d9b2-497b-9e4b-1cee94990697 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.823158] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263230, 'name': CreateVM_Task, 'duration_secs': 0.471609} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.823415] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 876.824137] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.824517] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.824852] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 876.825118] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71ee3b25-5abb-457d-b810-d7e88672b86b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.829446] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 876.829446] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52488d94-27bd-40c7-3331-3be1c20868db" [ 876.829446] env[62952]: _type = "Task" [ 876.829446] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.836833] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52488d94-27bd-40c7-3331-3be1c20868db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.966884] env[62952]: DEBUG nova.network.neutron [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Successfully updated port: 484ad815-dff6-400c-93dd-babc8b8b5c9c {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.021404] env[62952]: DEBUG oslo_concurrency.lockutils [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.022212] env[62952]: DEBUG nova.compute.manager [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 877.026737] env[62952]: DEBUG oslo_concurrency.lockutils [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.178s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.028821] env[62952]: INFO nova.compute.claims [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 877.169674] env[62952]: DEBUG nova.network.neutron [req-c7b7823b-7007-4092-a461-363e620dfb80 req-853693a0-77ae-412c-b098-ebb1fff28988 service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Updated VIF entry in instance network info cache for port 3423511f-8828-4fca-aec2-79e921a53579. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 877.170224] env[62952]: DEBUG nova.network.neutron [req-c7b7823b-7007-4092-a461-363e620dfb80 req-853693a0-77ae-412c-b098-ebb1fff28988 service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Updating instance_info_cache with network_info: [{"id": "3423511f-8828-4fca-aec2-79e921a53579", "address": "fa:16:3e:3b:bd:29", "network": {"id": "d865ccd7-b8a3-49d7-811f-6d4466bbbf45", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1549996301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4484b1cb78ff4e7baac4b9a5d1e63549", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a5569c57-a27b-4d4d-a519-784be105114f", "external-id": "nsx-vlan-transportzone-640", "segmentation_id": 640, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3423511f-88", "ovs_interfaceid": "3423511f-8828-4fca-aec2-79e921a53579", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.340545] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52488d94-27bd-40c7-3331-3be1c20868db, 'name': SearchDatastore_Task, 'duration_secs': 0.011412} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.340903] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.341625] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 877.341625] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.341625] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.341794] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 877.341999] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-419771f1-521e-4b68-aefb-deb7f71aae5f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.350531] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.350744] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 877.351502] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe85d94e-446f-4d18-84fe-e4d15baa8ede {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.356798] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 877.356798] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52060a0e-71e4-8c28-0e6f-ef865d03ca03" [ 877.356798] env[62952]: _type = "Task" [ 877.356798] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.364226] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52060a0e-71e4-8c28-0e6f-ef865d03ca03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.470964] env[62952]: DEBUG oslo_concurrency.lockutils [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquiring lock "refresh_cache-2b11ad9b-63d8-40da-8f68-3d8260c63bc7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.471183] env[62952]: DEBUG oslo_concurrency.lockutils [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquired lock "refresh_cache-2b11ad9b-63d8-40da-8f68-3d8260c63bc7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.471391] env[62952]: DEBUG nova.network.neutron [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 877.533610] env[62952]: DEBUG nova.compute.utils [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 877.537040] env[62952]: DEBUG nova.compute.manager [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 877.537230] env[62952]: DEBUG nova.network.neutron [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 877.577015] env[62952]: DEBUG nova.policy [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ffc122d64b19432aae8d9627d2928c06', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5080fbeac0744478843ab28efea8fc18', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 877.673067] env[62952]: DEBUG oslo_concurrency.lockutils [req-c7b7823b-7007-4092-a461-363e620dfb80 req-853693a0-77ae-412c-b098-ebb1fff28988 service nova] Releasing lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.869989] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52060a0e-71e4-8c28-0e6f-ef865d03ca03, 'name': SearchDatastore_Task, 'duration_secs': 0.008144} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.871221] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b55f0cd-9457-4946-9b6c-add25fac2da6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.875616] env[62952]: DEBUG nova.network.neutron [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Successfully created port: 5e2e9d1e-dce4-404e-bf41-1a7897be3758 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 877.878995] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 877.878995] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5217f0d4-db34-02f7-d948-c098916aa34d" [ 877.878995] env[62952]: _type = "Task" [ 877.878995] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.887947] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5217f0d4-db34-02f7-d948-c098916aa34d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.022737] env[62952]: DEBUG nova.network.neutron [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 878.037602] env[62952]: DEBUG nova.compute.manager [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 878.261083] env[62952]: DEBUG nova.network.neutron [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Updating instance_info_cache with network_info: [{"id": "484ad815-dff6-400c-93dd-babc8b8b5c9c", "address": "fa:16:3e:6c:d1:47", "network": {"id": "b8a4dc85-3768-40f0-bfb8-fcd3fa0e1aae", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1247193023-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c51c89d19ed4111a5c1631f0ad47cba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap484ad815-df", "ovs_interfaceid": "484ad815-dff6-400c-93dd-babc8b8b5c9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.392079] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5217f0d4-db34-02f7-d948-c098916aa34d, 'name': SearchDatastore_Task, 'duration_secs': 0.009059} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.393510] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.393510] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 67b729be-fb66-4ed1-bbea-e62216d460d5/67b729be-fb66-4ed1-bbea-e62216d460d5.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 878.393510] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2a896a25-be51-4976-a508-57c8ad8370aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.400050] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 878.400050] env[62952]: value = "task-1263231" [ 878.400050] env[62952]: _type = "Task" [ 878.400050] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.407643] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263231, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.436642] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693a560a-43ef-43e2-bc01-9f1ae0605533 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.444441] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e36521a-fc64-429b-924a-e579828887d4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.479261] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-447a7a66-8f04-4b49-aa91-efb46bc94d11 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.487232] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ee0f0b-d521-4397-a300-bf376a1181d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.495102] env[62952]: DEBUG nova.compute.manager [req-f01836f4-385f-4bb0-8679-a759ab79255a req-14a31870-cee2-4a5f-b416-21d68eed6580 service nova] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Received event network-vif-plugged-484ad815-dff6-400c-93dd-babc8b8b5c9c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 878.495489] env[62952]: DEBUG oslo_concurrency.lockutils [req-f01836f4-385f-4bb0-8679-a759ab79255a req-14a31870-cee2-4a5f-b416-21d68eed6580 service nova] Acquiring lock "2b11ad9b-63d8-40da-8f68-3d8260c63bc7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.495799] env[62952]: DEBUG oslo_concurrency.lockutils [req-f01836f4-385f-4bb0-8679-a759ab79255a req-14a31870-cee2-4a5f-b416-21d68eed6580 service nova] Lock "2b11ad9b-63d8-40da-8f68-3d8260c63bc7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.496071] env[62952]: DEBUG oslo_concurrency.lockutils [req-f01836f4-385f-4bb0-8679-a759ab79255a req-14a31870-cee2-4a5f-b416-21d68eed6580 service nova] Lock "2b11ad9b-63d8-40da-8f68-3d8260c63bc7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.496338] env[62952]: DEBUG nova.compute.manager [req-f01836f4-385f-4bb0-8679-a759ab79255a req-14a31870-cee2-4a5f-b416-21d68eed6580 service nova] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] No waiting events found dispatching network-vif-plugged-484ad815-dff6-400c-93dd-babc8b8b5c9c {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 878.496546] env[62952]: WARNING nova.compute.manager [req-f01836f4-385f-4bb0-8679-a759ab79255a req-14a31870-cee2-4a5f-b416-21d68eed6580 service nova] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Received unexpected event network-vif-plugged-484ad815-dff6-400c-93dd-babc8b8b5c9c for instance with vm_state building and task_state spawning. [ 878.496718] env[62952]: DEBUG nova.compute.manager [req-f01836f4-385f-4bb0-8679-a759ab79255a req-14a31870-cee2-4a5f-b416-21d68eed6580 service nova] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Received event network-changed-484ad815-dff6-400c-93dd-babc8b8b5c9c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 878.496890] env[62952]: DEBUG nova.compute.manager [req-f01836f4-385f-4bb0-8679-a759ab79255a req-14a31870-cee2-4a5f-b416-21d68eed6580 service nova] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Refreshing instance network info cache due to event network-changed-484ad815-dff6-400c-93dd-babc8b8b5c9c. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 878.497747] env[62952]: DEBUG oslo_concurrency.lockutils [req-f01836f4-385f-4bb0-8679-a759ab79255a req-14a31870-cee2-4a5f-b416-21d68eed6580 service nova] Acquiring lock "refresh_cache-2b11ad9b-63d8-40da-8f68-3d8260c63bc7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.506807] env[62952]: DEBUG nova.compute.provider_tree [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.763933] env[62952]: DEBUG oslo_concurrency.lockutils [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Releasing lock "refresh_cache-2b11ad9b-63d8-40da-8f68-3d8260c63bc7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.764334] env[62952]: DEBUG nova.compute.manager [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Instance network_info: |[{"id": "484ad815-dff6-400c-93dd-babc8b8b5c9c", "address": "fa:16:3e:6c:d1:47", "network": {"id": "b8a4dc85-3768-40f0-bfb8-fcd3fa0e1aae", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1247193023-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c51c89d19ed4111a5c1631f0ad47cba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap484ad815-df", "ovs_interfaceid": "484ad815-dff6-400c-93dd-babc8b8b5c9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 878.764718] env[62952]: DEBUG oslo_concurrency.lockutils [req-f01836f4-385f-4bb0-8679-a759ab79255a req-14a31870-cee2-4a5f-b416-21d68eed6580 service nova] Acquired lock "refresh_cache-2b11ad9b-63d8-40da-8f68-3d8260c63bc7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.764953] env[62952]: DEBUG nova.network.neutron [req-f01836f4-385f-4bb0-8679-a759ab79255a req-14a31870-cee2-4a5f-b416-21d68eed6580 service nova] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Refreshing network info cache for port 484ad815-dff6-400c-93dd-babc8b8b5c9c {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 878.766375] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:d1:47', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55c757ac-f8b2-466d-b634-07dbd100b312', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '484ad815-dff6-400c-93dd-babc8b8b5c9c', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.775015] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Creating folder: Project (0c51c89d19ed4111a5c1631f0ad47cba). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 878.776362] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6f50eaee-8700-47e2-9606-c349ce4d7daf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.790427] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Created folder: Project (0c51c89d19ed4111a5c1631f0ad47cba) in parent group-v271811. [ 878.790645] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Creating folder: Instances. Parent ref: group-v271863. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 878.791161] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6285c894-d340-483b-b583-4d67fe17fced {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.803512] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Created folder: Instances in parent group-v271863. [ 878.803921] env[62952]: DEBUG oslo.service.loopingcall [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.804201] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 878.804481] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-864d2476-d871-471e-b064-7ab53e2047bd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.827509] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 878.827509] env[62952]: value = "task-1263234" [ 878.827509] env[62952]: _type = "Task" [ 878.827509] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.837454] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263234, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.910997] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263231, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454726} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.911309] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 67b729be-fb66-4ed1-bbea-e62216d460d5/67b729be-fb66-4ed1-bbea-e62216d460d5.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 878.911533] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 878.911807] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f7e786d1-9a56-4432-a273-3d21ce006f27 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.919015] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 878.919015] env[62952]: value = "task-1263235" [ 878.919015] env[62952]: _type = "Task" [ 878.919015] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.927050] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263235, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.011045] env[62952]: DEBUG nova.scheduler.client.report [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.047455] env[62952]: DEBUG nova.compute.manager [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 879.066418] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "f906dc47-12cb-46f5-ae5c-53cfcc8765a8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.066418] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "f906dc47-12cb-46f5-ae5c-53cfcc8765a8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.066598] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "f906dc47-12cb-46f5-ae5c-53cfcc8765a8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.066711] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "f906dc47-12cb-46f5-ae5c-53cfcc8765a8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.066892] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "f906dc47-12cb-46f5-ae5c-53cfcc8765a8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.070818] env[62952]: INFO nova.compute.manager [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Terminating instance [ 879.073348] env[62952]: DEBUG nova.compute.manager [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 879.073505] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 879.074496] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5553bd76-3195-4acf-a322-408c8ea19668 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.079436] env[62952]: DEBUG nova.virt.hardware [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.079660] env[62952]: DEBUG nova.virt.hardware [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.079816] env[62952]: DEBUG nova.virt.hardware [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.079999] env[62952]: DEBUG nova.virt.hardware [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.080171] env[62952]: DEBUG nova.virt.hardware [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.080368] env[62952]: DEBUG nova.virt.hardware [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.080609] env[62952]: DEBUG nova.virt.hardware [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.080775] env[62952]: DEBUG nova.virt.hardware [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.080942] env[62952]: DEBUG nova.virt.hardware [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.081124] env[62952]: DEBUG nova.virt.hardware [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.081305] env[62952]: DEBUG nova.virt.hardware [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.082406] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7950de74-1b65-4fdc-9ea0-4235fd83ea9b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.092221] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae45eac3-a184-4963-8b10-9ae1f63555fb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.098223] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 879.098470] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c54ceb57-a137-4b0b-8caa-afc021a69f17 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.111941] env[62952]: DEBUG oslo_vmware.api [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 879.111941] env[62952]: value = "task-1263236" [ 879.111941] env[62952]: _type = "Task" [ 879.111941] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.120382] env[62952]: DEBUG oslo_vmware.api [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263236, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.338029] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263234, 'name': CreateVM_Task, 'duration_secs': 0.337592} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.338029] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 879.338638] env[62952]: DEBUG oslo_concurrency.lockutils [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.338807] env[62952]: DEBUG oslo_concurrency.lockutils [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.339154] env[62952]: DEBUG oslo_concurrency.lockutils [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 879.339414] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-890e8253-46bb-4beb-9225-846ca7428323 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.345080] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Waiting for the task: (returnval){ [ 879.345080] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e38ed3-2748-e6b1-c333-58a507d483c8" [ 879.345080] env[62952]: _type = "Task" [ 879.345080] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.351984] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e38ed3-2748-e6b1-c333-58a507d483c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.428273] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263235, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060062} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.428627] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.429437] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7493c706-da6b-4079-aece-1b6092f2513d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.452833] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 67b729be-fb66-4ed1-bbea-e62216d460d5/67b729be-fb66-4ed1-bbea-e62216d460d5.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.455853] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0dbd641b-3784-4070-a9de-2e9c281978a3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.475598] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 879.475598] env[62952]: value = "task-1263237" [ 879.475598] env[62952]: _type = "Task" [ 879.475598] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.486019] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263237, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.517891] env[62952]: DEBUG oslo_concurrency.lockutils [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.491s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.518517] env[62952]: DEBUG nova.compute.manager [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 879.521147] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.530s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.522589] env[62952]: INFO nova.compute.claims [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.620970] env[62952]: DEBUG oslo_vmware.api [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263236, 'name': PowerOffVM_Task, 'duration_secs': 0.345752} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.621267] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 879.621446] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 879.621696] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-896ff580-f314-4139-803c-494f1231819b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.641707] env[62952]: DEBUG nova.network.neutron [req-f01836f4-385f-4bb0-8679-a759ab79255a req-14a31870-cee2-4a5f-b416-21d68eed6580 service nova] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Updated VIF entry in instance network info cache for port 484ad815-dff6-400c-93dd-babc8b8b5c9c. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 879.642076] env[62952]: DEBUG nova.network.neutron [req-f01836f4-385f-4bb0-8679-a759ab79255a req-14a31870-cee2-4a5f-b416-21d68eed6580 service nova] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Updating instance_info_cache with network_info: [{"id": "484ad815-dff6-400c-93dd-babc8b8b5c9c", "address": "fa:16:3e:6c:d1:47", "network": {"id": "b8a4dc85-3768-40f0-bfb8-fcd3fa0e1aae", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1247193023-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c51c89d19ed4111a5c1631f0ad47cba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55c757ac-f8b2-466d-b634-07dbd100b312", "external-id": "nsx-vlan-transportzone-159", "segmentation_id": 159, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap484ad815-df", "ovs_interfaceid": "484ad815-dff6-400c-93dd-babc8b8b5c9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.660972] env[62952]: DEBUG nova.network.neutron [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Successfully updated port: 5e2e9d1e-dce4-404e-bf41-1a7897be3758 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 879.718875] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 879.719118] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 879.719310] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Deleting the datastore file [datastore2] f906dc47-12cb-46f5-ae5c-53cfcc8765a8 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 879.719589] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3f86bdd6-b341-476b-b3ee-edd75bbb72cb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.727013] env[62952]: DEBUG oslo_vmware.api [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 879.727013] env[62952]: value = "task-1263239" [ 879.727013] env[62952]: _type = "Task" [ 879.727013] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.734948] env[62952]: DEBUG oslo_vmware.api [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263239, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.805099] env[62952]: DEBUG oslo_concurrency.lockutils [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.805369] env[62952]: DEBUG oslo_concurrency.lockutils [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.854994] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e38ed3-2748-e6b1-c333-58a507d483c8, 'name': SearchDatastore_Task, 'duration_secs': 0.040324} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.855402] env[62952]: DEBUG oslo_concurrency.lockutils [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.855701] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 879.856260] env[62952]: DEBUG oslo_concurrency.lockutils [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.856260] env[62952]: DEBUG oslo_concurrency.lockutils [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.856532] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.856891] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f79fb4f-2e22-4061-af8c-019c10319274 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.873681] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.874069] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 879.874705] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c182a259-995f-43b4-ab3b-5aad1b8df71e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.880849] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Waiting for the task: (returnval){ [ 879.880849] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5251d554-20b4-3ef1-29be-df7bdf1fce56" [ 879.880849] env[62952]: _type = "Task" [ 879.880849] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.888819] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5251d554-20b4-3ef1-29be-df7bdf1fce56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.985765] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.027464] env[62952]: DEBUG nova.compute.utils [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 880.031330] env[62952]: DEBUG nova.compute.manager [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 880.031445] env[62952]: DEBUG nova.network.neutron [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 880.081250] env[62952]: DEBUG nova.policy [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7eadc05f193645e6b6e86e9cdbbbaa71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '076fb3c53420447caf047f8422435152', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 880.145591] env[62952]: DEBUG oslo_concurrency.lockutils [req-f01836f4-385f-4bb0-8679-a759ab79255a req-14a31870-cee2-4a5f-b416-21d68eed6580 service nova] Releasing lock "refresh_cache-2b11ad9b-63d8-40da-8f68-3d8260c63bc7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.167032] env[62952]: DEBUG oslo_concurrency.lockutils [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.167157] env[62952]: DEBUG oslo_concurrency.lockutils [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.167487] env[62952]: DEBUG nova.network.neutron [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 880.237990] env[62952]: DEBUG oslo_vmware.api [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263239, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.309180] env[62952]: DEBUG nova.compute.utils [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 880.336086] env[62952]: DEBUG nova.network.neutron [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Successfully created port: dfcafefe-49c2-461d-b13c-9186be103ed0 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 880.392225] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5251d554-20b4-3ef1-29be-df7bdf1fce56, 'name': SearchDatastore_Task, 'duration_secs': 0.034782} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.393116] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ec45411-6c50-46f3-8383-81a6abc137b0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.398524] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Waiting for the task: (returnval){ [ 880.398524] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5257225b-4d59-065b-01a4-993a50e67edb" [ 880.398524] env[62952]: _type = "Task" [ 880.398524] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.405887] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5257225b-4d59-065b-01a4-993a50e67edb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.487545] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263237, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.537449] env[62952]: DEBUG nova.compute.manager [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 880.654697] env[62952]: DEBUG nova.compute.manager [req-acc7c7cf-d9fd-46ae-bf14-9baac0cf72a4 req-963c3b76-0771-4d5c-ba71-16bfe3f74169 service nova] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Received event network-vif-plugged-5e2e9d1e-dce4-404e-bf41-1a7897be3758 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 880.654934] env[62952]: DEBUG oslo_concurrency.lockutils [req-acc7c7cf-d9fd-46ae-bf14-9baac0cf72a4 req-963c3b76-0771-4d5c-ba71-16bfe3f74169 service nova] Acquiring lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.655128] env[62952]: DEBUG oslo_concurrency.lockutils [req-acc7c7cf-d9fd-46ae-bf14-9baac0cf72a4 req-963c3b76-0771-4d5c-ba71-16bfe3f74169 service nova] Lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.655301] env[62952]: DEBUG oslo_concurrency.lockutils [req-acc7c7cf-d9fd-46ae-bf14-9baac0cf72a4 req-963c3b76-0771-4d5c-ba71-16bfe3f74169 service nova] Lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.655477] env[62952]: DEBUG nova.compute.manager [req-acc7c7cf-d9fd-46ae-bf14-9baac0cf72a4 req-963c3b76-0771-4d5c-ba71-16bfe3f74169 service nova] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] No waiting events found dispatching network-vif-plugged-5e2e9d1e-dce4-404e-bf41-1a7897be3758 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 880.655708] env[62952]: WARNING nova.compute.manager [req-acc7c7cf-d9fd-46ae-bf14-9baac0cf72a4 req-963c3b76-0771-4d5c-ba71-16bfe3f74169 service nova] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Received unexpected event network-vif-plugged-5e2e9d1e-dce4-404e-bf41-1a7897be3758 for instance with vm_state building and task_state spawning. [ 880.655881] env[62952]: DEBUG nova.compute.manager [req-acc7c7cf-d9fd-46ae-bf14-9baac0cf72a4 req-963c3b76-0771-4d5c-ba71-16bfe3f74169 service nova] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Received event network-changed-5e2e9d1e-dce4-404e-bf41-1a7897be3758 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 880.656056] env[62952]: DEBUG nova.compute.manager [req-acc7c7cf-d9fd-46ae-bf14-9baac0cf72a4 req-963c3b76-0771-4d5c-ba71-16bfe3f74169 service nova] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Refreshing instance network info cache due to event network-changed-5e2e9d1e-dce4-404e-bf41-1a7897be3758. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 880.656260] env[62952]: DEBUG oslo_concurrency.lockutils [req-acc7c7cf-d9fd-46ae-bf14-9baac0cf72a4 req-963c3b76-0771-4d5c-ba71-16bfe3f74169 service nova] Acquiring lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.698563] env[62952]: DEBUG nova.network.neutron [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 880.736773] env[62952]: DEBUG oslo_vmware.api [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263239, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.525235} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.739539] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 880.739615] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 880.739829] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 880.740097] env[62952]: INFO nova.compute.manager [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Took 1.67 seconds to destroy the instance on the hypervisor. [ 880.740425] env[62952]: DEBUG oslo.service.loopingcall [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.742807] env[62952]: DEBUG nova.compute.manager [-] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 880.742978] env[62952]: DEBUG nova.network.neutron [-] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 880.815975] env[62952]: DEBUG oslo_concurrency.lockutils [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.886403] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d66eefa-14ba-4919-b5e2-e0fa63f1b27c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.894076] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4564377-6bae-4730-80bc-57137f3c8eaa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.933185] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e6fbd8-0f26-4b55-b128-45a6f7f1ff0b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.942020] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5257225b-4d59-065b-01a4-993a50e67edb, 'name': SearchDatastore_Task, 'duration_secs': 0.016061} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.942020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.942153] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 2b11ad9b-63d8-40da-8f68-3d8260c63bc7/2b11ad9b-63d8-40da-8f68-3d8260c63bc7.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 880.942395] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bc03c5ac-cee3-498e-aac6-bf67e250ee9b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.946172] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba16565-ce2b-49fb-887f-a00e442e740e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.960035] env[62952]: DEBUG nova.compute.provider_tree [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.962989] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Waiting for the task: (returnval){ [ 880.962989] env[62952]: value = "task-1263240" [ 880.962989] env[62952]: _type = "Task" [ 880.962989] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.975286] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': task-1263240, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.985996] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263237, 'name': ReconfigVM_Task, 'duration_secs': 1.047006} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.986345] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 67b729be-fb66-4ed1-bbea-e62216d460d5/67b729be-fb66-4ed1-bbea-e62216d460d5.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.986999] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2a619b59-27d4-44c4-9d68-2faf64cab0a1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.996640] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 880.996640] env[62952]: value = "task-1263241" [ 880.996640] env[62952]: _type = "Task" [ 880.996640] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.004911] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263241, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.092188] env[62952]: DEBUG nova.network.neutron [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance_info_cache with network_info: [{"id": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "address": "fa:16:3e:eb:c2:8d", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e2e9d1e-dc", "ovs_interfaceid": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.464600] env[62952]: DEBUG nova.scheduler.client.report [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 881.478109] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': task-1263240, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.506926] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263241, 'name': Rename_Task, 'duration_secs': 0.160231} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.507255] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 881.507408] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f450272c-4f69-485b-b6c1-1b6da83aa9a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.513939] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 881.513939] env[62952]: value = "task-1263242" [ 881.513939] env[62952]: _type = "Task" [ 881.513939] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.521643] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263242, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.546786] env[62952]: DEBUG nova.compute.manager [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 881.577731] env[62952]: DEBUG nova.virt.hardware [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 881.578095] env[62952]: DEBUG nova.virt.hardware [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 881.578211] env[62952]: DEBUG nova.virt.hardware [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 881.578356] env[62952]: DEBUG nova.virt.hardware [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 881.578507] env[62952]: DEBUG nova.virt.hardware [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 881.578916] env[62952]: DEBUG nova.virt.hardware [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 881.578916] env[62952]: DEBUG nova.virt.hardware [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 881.579279] env[62952]: DEBUG nova.virt.hardware [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 881.579279] env[62952]: DEBUG nova.virt.hardware [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 881.579389] env[62952]: DEBUG nova.virt.hardware [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 881.579564] env[62952]: DEBUG nova.virt.hardware [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 881.580480] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03079503-afac-4ab6-84c1-3f3020ab7bcc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.588870] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714b17b2-58be-47c7-8b9b-181d95ed2f3c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.595753] env[62952]: DEBUG oslo_concurrency.lockutils [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.595753] env[62952]: DEBUG nova.compute.manager [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Instance network_info: |[{"id": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "address": "fa:16:3e:eb:c2:8d", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e2e9d1e-dc", "ovs_interfaceid": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 881.602492] env[62952]: DEBUG oslo_concurrency.lockutils [req-acc7c7cf-d9fd-46ae-bf14-9baac0cf72a4 req-963c3b76-0771-4d5c-ba71-16bfe3f74169 service nova] Acquired lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.602702] env[62952]: DEBUG nova.network.neutron [req-acc7c7cf-d9fd-46ae-bf14-9baac0cf72a4 req-963c3b76-0771-4d5c-ba71-16bfe3f74169 service nova] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Refreshing network info cache for port 5e2e9d1e-dce4-404e-bf41-1a7897be3758 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 881.603950] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:eb:c2:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e1c9bb98-73a9-48eb-856e-a541afe9b07b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5e2e9d1e-dce4-404e-bf41-1a7897be3758', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 881.617857] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Creating folder: Project (5080fbeac0744478843ab28efea8fc18). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 881.619902] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4f1b7fa5-ab0e-4fbb-88c6-de72866f93b3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.630575] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Created folder: Project (5080fbeac0744478843ab28efea8fc18) in parent group-v271811. [ 881.630794] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Creating folder: Instances. Parent ref: group-v271866. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 881.631044] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5fa2c340-e1ac-4880-84fb-c39a5d2f9d36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.639893] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Created folder: Instances in parent group-v271866. [ 881.640135] env[62952]: DEBUG oslo.service.loopingcall [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.640328] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 881.640575] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6dc587b6-4b74-41ed-a962-52c560e67891 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.662964] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 881.662964] env[62952]: value = "task-1263245" [ 881.662964] env[62952]: _type = "Task" [ 881.662964] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.673350] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263245, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.758454] env[62952]: DEBUG nova.network.neutron [-] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.849582] env[62952]: DEBUG nova.network.neutron [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Successfully updated port: dfcafefe-49c2-461d-b13c-9186be103ed0 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 881.913447] env[62952]: DEBUG oslo_concurrency.lockutils [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.913821] env[62952]: DEBUG oslo_concurrency.lockutils [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.914110] env[62952]: INFO nova.compute.manager [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Attaching volume 42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d to /dev/sdb [ 881.949278] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea53e86-34d0-4d9b-a49d-c436726c5d4f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.956729] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8518424-256e-4228-8b0a-2d0fc3410d33 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.970483] env[62952]: DEBUG nova.virt.block_device [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Updating existing volume attachment record: 2af238dd-b9bf-4be1-bf27-311b5b58ec7d {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 881.975844] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.976343] env[62952]: DEBUG nova.compute.manager [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 881.978913] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.880s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.980537] env[62952]: INFO nova.compute.claims [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 881.988079] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': task-1263240, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.590662} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.988079] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 2b11ad9b-63d8-40da-8f68-3d8260c63bc7/2b11ad9b-63d8-40da-8f68-3d8260c63bc7.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 881.988196] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.988406] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-60177e68-c6b6-49ba-a6c2-920a69b0d035 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.998599] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Waiting for the task: (returnval){ [ 881.998599] env[62952]: value = "task-1263246" [ 881.998599] env[62952]: _type = "Task" [ 881.998599] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.006183] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': task-1263246, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.025554] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263242, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.172774] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263245, 'name': CreateVM_Task, 'duration_secs': 0.499777} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.172949] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 882.173717] env[62952]: DEBUG oslo_concurrency.lockutils [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.173994] env[62952]: DEBUG oslo_concurrency.lockutils [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.174391] env[62952]: DEBUG oslo_concurrency.lockutils [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 882.174703] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a80a25b-eec9-438f-8db1-72d7e4172552 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.179113] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 882.179113] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5233feaa-3b97-1a80-7049-180c7f0f6cab" [ 882.179113] env[62952]: _type = "Task" [ 882.179113] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.186959] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5233feaa-3b97-1a80-7049-180c7f0f6cab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.259686] env[62952]: INFO nova.compute.manager [-] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Took 1.52 seconds to deallocate network for instance. [ 882.342274] env[62952]: DEBUG nova.network.neutron [req-acc7c7cf-d9fd-46ae-bf14-9baac0cf72a4 req-963c3b76-0771-4d5c-ba71-16bfe3f74169 service nova] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updated VIF entry in instance network info cache for port 5e2e9d1e-dce4-404e-bf41-1a7897be3758. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 882.342687] env[62952]: DEBUG nova.network.neutron [req-acc7c7cf-d9fd-46ae-bf14-9baac0cf72a4 req-963c3b76-0771-4d5c-ba71-16bfe3f74169 service nova] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance_info_cache with network_info: [{"id": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "address": "fa:16:3e:eb:c2:8d", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e2e9d1e-dc", "ovs_interfaceid": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.351901] env[62952]: DEBUG oslo_concurrency.lockutils [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "refresh_cache-dd700ee9-0957-4210-be50-e4c8ac3ca456" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.352061] env[62952]: DEBUG oslo_concurrency.lockutils [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "refresh_cache-dd700ee9-0957-4210-be50-e4c8ac3ca456" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.352209] env[62952]: DEBUG nova.network.neutron [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 882.484731] env[62952]: DEBUG nova.compute.utils [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 882.488699] env[62952]: DEBUG nova.compute.manager [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 882.488877] env[62952]: DEBUG nova.network.neutron [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 882.511127] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': task-1263246, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07719} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.511434] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 882.512172] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f0ed10-8677-4c17-b909-c1f58023bf96 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.543334] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 2b11ad9b-63d8-40da-8f68-3d8260c63bc7/2b11ad9b-63d8-40da-8f68-3d8260c63bc7.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.543744] env[62952]: DEBUG oslo_vmware.api [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263242, 'name': PowerOnVM_Task, 'duration_secs': 0.990231} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.545303] env[62952]: DEBUG nova.policy [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3ec1aecc0cf64aaea875a4280b114874', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fbb5ccb4d615429facd6932fb55b04e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 882.546887] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86d76602-c71c-4649-be12-0dd8fddd25be {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.561530] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 882.561798] env[62952]: INFO nova.compute.manager [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Took 8.34 seconds to spawn the instance on the hypervisor. [ 882.561994] env[62952]: DEBUG nova.compute.manager [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.564897] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d5b6e7-574f-43ba-8381-36069838e0fd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.576341] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Waiting for the task: (returnval){ [ 882.576341] env[62952]: value = "task-1263250" [ 882.576341] env[62952]: _type = "Task" [ 882.576341] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.585351] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': task-1263250, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.687209] env[62952]: DEBUG nova.compute.manager [req-e69be5bc-2b8e-421d-8c25-5eefaf3261dd req-174c1a15-77eb-4079-a360-7e7fcaef1ae8 service nova] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Received event network-vif-deleted-5805101a-366b-4401-9db9-f69ed60ff18d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 882.687424] env[62952]: DEBUG nova.compute.manager [req-e69be5bc-2b8e-421d-8c25-5eefaf3261dd req-174c1a15-77eb-4079-a360-7e7fcaef1ae8 service nova] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Received event network-vif-plugged-dfcafefe-49c2-461d-b13c-9186be103ed0 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 882.687627] env[62952]: DEBUG oslo_concurrency.lockutils [req-e69be5bc-2b8e-421d-8c25-5eefaf3261dd req-174c1a15-77eb-4079-a360-7e7fcaef1ae8 service nova] Acquiring lock "dd700ee9-0957-4210-be50-e4c8ac3ca456-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.687960] env[62952]: DEBUG oslo_concurrency.lockutils [req-e69be5bc-2b8e-421d-8c25-5eefaf3261dd req-174c1a15-77eb-4079-a360-7e7fcaef1ae8 service nova] Lock "dd700ee9-0957-4210-be50-e4c8ac3ca456-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.688194] env[62952]: DEBUG oslo_concurrency.lockutils [req-e69be5bc-2b8e-421d-8c25-5eefaf3261dd req-174c1a15-77eb-4079-a360-7e7fcaef1ae8 service nova] Lock "dd700ee9-0957-4210-be50-e4c8ac3ca456-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.688333] env[62952]: DEBUG nova.compute.manager [req-e69be5bc-2b8e-421d-8c25-5eefaf3261dd req-174c1a15-77eb-4079-a360-7e7fcaef1ae8 service nova] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] No waiting events found dispatching network-vif-plugged-dfcafefe-49c2-461d-b13c-9186be103ed0 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 882.688553] env[62952]: WARNING nova.compute.manager [req-e69be5bc-2b8e-421d-8c25-5eefaf3261dd req-174c1a15-77eb-4079-a360-7e7fcaef1ae8 service nova] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Received unexpected event network-vif-plugged-dfcafefe-49c2-461d-b13c-9186be103ed0 for instance with vm_state building and task_state spawning. [ 882.688728] env[62952]: DEBUG nova.compute.manager [req-e69be5bc-2b8e-421d-8c25-5eefaf3261dd req-174c1a15-77eb-4079-a360-7e7fcaef1ae8 service nova] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Received event network-changed-dfcafefe-49c2-461d-b13c-9186be103ed0 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 882.688936] env[62952]: DEBUG nova.compute.manager [req-e69be5bc-2b8e-421d-8c25-5eefaf3261dd req-174c1a15-77eb-4079-a360-7e7fcaef1ae8 service nova] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Refreshing instance network info cache due to event network-changed-dfcafefe-49c2-461d-b13c-9186be103ed0. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 882.689136] env[62952]: DEBUG oslo_concurrency.lockutils [req-e69be5bc-2b8e-421d-8c25-5eefaf3261dd req-174c1a15-77eb-4079-a360-7e7fcaef1ae8 service nova] Acquiring lock "refresh_cache-dd700ee9-0957-4210-be50-e4c8ac3ca456" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.694528] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5233feaa-3b97-1a80-7049-180c7f0f6cab, 'name': SearchDatastore_Task, 'duration_secs': 0.018956} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.694830] env[62952]: DEBUG oslo_concurrency.lockutils [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.695123] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 882.695367] env[62952]: DEBUG oslo_concurrency.lockutils [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.695537] env[62952]: DEBUG oslo_concurrency.lockutils [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.695722] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.696082] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-afab0d7f-d254-4848-9499-996d0dc015bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.706679] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.706878] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 882.707685] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-727318de-7b63-42cd-930b-4660c21cc0a9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.713548] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 882.713548] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52cf6345-13b9-5ffc-ee67-649a439d29f8" [ 882.713548] env[62952]: _type = "Task" [ 882.713548] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.721274] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52cf6345-13b9-5ffc-ee67-649a439d29f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.766738] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.843421] env[62952]: DEBUG nova.network.neutron [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Successfully created port: d4ed51c5-2a91-40ed-8ce9-62f6109ab3b8 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 882.845702] env[62952]: DEBUG oslo_concurrency.lockutils [req-acc7c7cf-d9fd-46ae-bf14-9baac0cf72a4 req-963c3b76-0771-4d5c-ba71-16bfe3f74169 service nova] Releasing lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.902541] env[62952]: DEBUG nova.network.neutron [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 882.993611] env[62952]: DEBUG nova.compute.manager [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 883.091177] env[62952]: INFO nova.compute.manager [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Took 38.78 seconds to build instance. [ 883.096047] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': task-1263250, 'name': ReconfigVM_Task, 'duration_secs': 0.421929} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.096536] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 2b11ad9b-63d8-40da-8f68-3d8260c63bc7/2b11ad9b-63d8-40da-8f68-3d8260c63bc7.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.097175] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a4c8f96c-b110-48fd-8356-37a0f97a8d8d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.111405] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Waiting for the task: (returnval){ [ 883.111405] env[62952]: value = "task-1263251" [ 883.111405] env[62952]: _type = "Task" [ 883.111405] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.125248] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': task-1263251, 'name': Rename_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.130512] env[62952]: DEBUG nova.network.neutron [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Updating instance_info_cache with network_info: [{"id": "dfcafefe-49c2-461d-b13c-9186be103ed0", "address": "fa:16:3e:a8:b8:33", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfcafefe-49", "ovs_interfaceid": "dfcafefe-49c2-461d-b13c-9186be103ed0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.225943] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52cf6345-13b9-5ffc-ee67-649a439d29f8, 'name': SearchDatastore_Task, 'duration_secs': 0.017288} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.226788] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38a6f8f1-de72-48e2-ba38-62aceae7352d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.232221] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 883.232221] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]529a6150-5796-034f-a1c6-db29acfc452a" [ 883.232221] env[62952]: _type = "Task" [ 883.232221] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.240156] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]529a6150-5796-034f-a1c6-db29acfc452a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.387030] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b8c206-19c5-47c7-87d3-41c57ba8bb2d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.396556] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95fb30e5-41c5-4304-96c8-345750ed2d5a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.428488] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bed4b7a-76c2-421a-bb3f-f30bb88d67b1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.439224] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47920bd5-3f43-4526-b65a-6939d5c540a0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.453616] env[62952]: DEBUG nova.compute.provider_tree [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.597053] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1160204-1061-4813-9496-8e6e1f48ccb9 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Lock "67b729be-fb66-4ed1-bbea-e62216d460d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.731s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.621907] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': task-1263251, 'name': Rename_Task, 'duration_secs': 0.247991} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.622192] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 883.622430] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-874145a5-b396-4ae1-918a-ed937546813e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.628359] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Waiting for the task: (returnval){ [ 883.628359] env[62952]: value = "task-1263252" [ 883.628359] env[62952]: _type = "Task" [ 883.628359] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.635504] env[62952]: DEBUG oslo_concurrency.lockutils [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "refresh_cache-dd700ee9-0957-4210-be50-e4c8ac3ca456" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.635788] env[62952]: DEBUG nova.compute.manager [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Instance network_info: |[{"id": "dfcafefe-49c2-461d-b13c-9186be103ed0", "address": "fa:16:3e:a8:b8:33", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfcafefe-49", "ovs_interfaceid": "dfcafefe-49c2-461d-b13c-9186be103ed0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 883.636048] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': task-1263252, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.636280] env[62952]: DEBUG oslo_concurrency.lockutils [req-e69be5bc-2b8e-421d-8c25-5eefaf3261dd req-174c1a15-77eb-4079-a360-7e7fcaef1ae8 service nova] Acquired lock "refresh_cache-dd700ee9-0957-4210-be50-e4c8ac3ca456" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.636455] env[62952]: DEBUG nova.network.neutron [req-e69be5bc-2b8e-421d-8c25-5eefaf3261dd req-174c1a15-77eb-4079-a360-7e7fcaef1ae8 service nova] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Refreshing network info cache for port dfcafefe-49c2-461d-b13c-9186be103ed0 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 883.637505] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a8:b8:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '365ac5b1-6d83-4dfe-887f-60574d7f6124', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dfcafefe-49c2-461d-b13c-9186be103ed0', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 883.644810] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Creating folder: Project (076fb3c53420447caf047f8422435152). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 883.645679] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-73a46522-98ee-4de8-b84c-ee789c81470c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.655830] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Created folder: Project (076fb3c53420447caf047f8422435152) in parent group-v271811. [ 883.656024] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Creating folder: Instances. Parent ref: group-v271871. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 883.656239] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-817fe2e5-7c44-4f46-8d26-f2749242d9c1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.664697] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Created folder: Instances in parent group-v271871. [ 883.664924] env[62952]: DEBUG oslo.service.loopingcall [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.665123] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 883.665315] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bb53e78c-a843-4798-80b9-1f8ded90e70e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.684283] env[62952]: INFO nova.compute.manager [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Rescuing [ 883.684449] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Acquiring lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.684600] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Acquired lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.684762] env[62952]: DEBUG nova.network.neutron [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 883.686855] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 883.686855] env[62952]: value = "task-1263255" [ 883.686855] env[62952]: _type = "Task" [ 883.686855] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.695376] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263255, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.743323] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]529a6150-5796-034f-a1c6-db29acfc452a, 'name': SearchDatastore_Task, 'duration_secs': 0.010932} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.743626] env[62952]: DEBUG oslo_concurrency.lockutils [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.743912] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 6e289c20-8dd2-4680-b816-9bbe82fe7aec/6e289c20-8dd2-4680-b816-9bbe82fe7aec.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 883.744206] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7845a014-f40d-4669-9fc0-90197dbd919e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.751514] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 883.751514] env[62952]: value = "task-1263256" [ 883.751514] env[62952]: _type = "Task" [ 883.751514] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.759993] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263256, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.957907] env[62952]: DEBUG nova.scheduler.client.report [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.005595] env[62952]: DEBUG nova.compute.manager [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 884.038709] env[62952]: DEBUG nova.virt.hardware [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 884.039066] env[62952]: DEBUG nova.virt.hardware [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 884.039270] env[62952]: DEBUG nova.virt.hardware [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.039516] env[62952]: DEBUG nova.virt.hardware [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 884.039695] env[62952]: DEBUG nova.virt.hardware [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.039871] env[62952]: DEBUG nova.virt.hardware [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 884.040185] env[62952]: DEBUG nova.virt.hardware [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 884.040447] env[62952]: DEBUG nova.virt.hardware [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 884.040711] env[62952]: DEBUG nova.virt.hardware [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 884.040949] env[62952]: DEBUG nova.virt.hardware [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 884.041187] env[62952]: DEBUG nova.virt.hardware [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.042235] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb67e998-f31d-42c0-adb2-bc0b883c9540 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.053184] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7281ad53-c632-4bc1-b21e-e8d944e491a9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.099838] env[62952]: DEBUG nova.compute.manager [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 884.142252] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': task-1263252, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.201053] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263255, 'name': CreateVM_Task, 'duration_secs': 0.3952} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.201310] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 884.202091] env[62952]: DEBUG oslo_concurrency.lockutils [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.202272] env[62952]: DEBUG oslo_concurrency.lockutils [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.202657] env[62952]: DEBUG oslo_concurrency.lockutils [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 884.202944] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37bbff41-6c12-42b1-8b91-b9a00e4d8259 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.211347] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 884.211347] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e62a7e-4cb0-1995-9826-de9c8290e280" [ 884.211347] env[62952]: _type = "Task" [ 884.211347] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.224290] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e62a7e-4cb0-1995-9826-de9c8290e280, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.266049] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263256, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.463948] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.469444] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.559s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.469444] env[62952]: INFO nova.compute.claims [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 884.531029] env[62952]: DEBUG nova.compute.manager [req-5bdefe4f-c6ba-4f49-b776-1130444685b3 req-2a7fc7d4-ae80-443b-b9e5-09bb99d4477b service nova] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Received event network-vif-plugged-d4ed51c5-2a91-40ed-8ce9-62f6109ab3b8 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 884.531246] env[62952]: DEBUG oslo_concurrency.lockutils [req-5bdefe4f-c6ba-4f49-b776-1130444685b3 req-2a7fc7d4-ae80-443b-b9e5-09bb99d4477b service nova] Acquiring lock "3416eb38-b961-4d48-8b37-8b7b92375dc9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.531657] env[62952]: DEBUG oslo_concurrency.lockutils [req-5bdefe4f-c6ba-4f49-b776-1130444685b3 req-2a7fc7d4-ae80-443b-b9e5-09bb99d4477b service nova] Lock "3416eb38-b961-4d48-8b37-8b7b92375dc9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.531657] env[62952]: DEBUG oslo_concurrency.lockutils [req-5bdefe4f-c6ba-4f49-b776-1130444685b3 req-2a7fc7d4-ae80-443b-b9e5-09bb99d4477b service nova] Lock "3416eb38-b961-4d48-8b37-8b7b92375dc9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.531933] env[62952]: DEBUG nova.compute.manager [req-5bdefe4f-c6ba-4f49-b776-1130444685b3 req-2a7fc7d4-ae80-443b-b9e5-09bb99d4477b service nova] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] No waiting events found dispatching network-vif-plugged-d4ed51c5-2a91-40ed-8ce9-62f6109ab3b8 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 884.532140] env[62952]: WARNING nova.compute.manager [req-5bdefe4f-c6ba-4f49-b776-1130444685b3 req-2a7fc7d4-ae80-443b-b9e5-09bb99d4477b service nova] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Received unexpected event network-vif-plugged-d4ed51c5-2a91-40ed-8ce9-62f6109ab3b8 for instance with vm_state building and task_state spawning. [ 884.596478] env[62952]: DEBUG nova.network.neutron [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Successfully updated port: d4ed51c5-2a91-40ed-8ce9-62f6109ab3b8 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 884.609165] env[62952]: DEBUG nova.network.neutron [req-e69be5bc-2b8e-421d-8c25-5eefaf3261dd req-174c1a15-77eb-4079-a360-7e7fcaef1ae8 service nova] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Updated VIF entry in instance network info cache for port dfcafefe-49c2-461d-b13c-9186be103ed0. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 884.610085] env[62952]: DEBUG nova.network.neutron [req-e69be5bc-2b8e-421d-8c25-5eefaf3261dd req-174c1a15-77eb-4079-a360-7e7fcaef1ae8 service nova] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Updating instance_info_cache with network_info: [{"id": "dfcafefe-49c2-461d-b13c-9186be103ed0", "address": "fa:16:3e:a8:b8:33", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfcafefe-49", "ovs_interfaceid": "dfcafefe-49c2-461d-b13c-9186be103ed0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.639889] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.640904] env[62952]: DEBUG nova.network.neutron [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Updating instance_info_cache with network_info: [{"id": "3423511f-8828-4fca-aec2-79e921a53579", "address": "fa:16:3e:3b:bd:29", "network": {"id": "d865ccd7-b8a3-49d7-811f-6d4466bbbf45", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1549996301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4484b1cb78ff4e7baac4b9a5d1e63549", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a5569c57-a27b-4d4d-a519-784be105114f", "external-id": "nsx-vlan-transportzone-640", "segmentation_id": 640, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3423511f-88", "ovs_interfaceid": "3423511f-8828-4fca-aec2-79e921a53579", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.645392] env[62952]: DEBUG oslo_vmware.api [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': task-1263252, 'name': PowerOnVM_Task, 'duration_secs': 0.612129} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.645851] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 884.646086] env[62952]: INFO nova.compute.manager [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Took 8.02 seconds to spawn the instance on the hypervisor. [ 884.646280] env[62952]: DEBUG nova.compute.manager [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.647098] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a1e14db-17a4-473a-859a-f4c950f84a7d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.721390] env[62952]: DEBUG nova.compute.manager [req-62637716-10c8-404a-a986-cc58e5a1c33b req-c34f670c-8b93-47d2-abb1-e68d111de230 service nova] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Received event network-changed-d4ed51c5-2a91-40ed-8ce9-62f6109ab3b8 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 884.721849] env[62952]: DEBUG nova.compute.manager [req-62637716-10c8-404a-a986-cc58e5a1c33b req-c34f670c-8b93-47d2-abb1-e68d111de230 service nova] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Refreshing instance network info cache due to event network-changed-d4ed51c5-2a91-40ed-8ce9-62f6109ab3b8. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 884.722299] env[62952]: DEBUG oslo_concurrency.lockutils [req-62637716-10c8-404a-a986-cc58e5a1c33b req-c34f670c-8b93-47d2-abb1-e68d111de230 service nova] Acquiring lock "refresh_cache-3416eb38-b961-4d48-8b37-8b7b92375dc9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.722728] env[62952]: DEBUG oslo_concurrency.lockutils [req-62637716-10c8-404a-a986-cc58e5a1c33b req-c34f670c-8b93-47d2-abb1-e68d111de230 service nova] Acquired lock "refresh_cache-3416eb38-b961-4d48-8b37-8b7b92375dc9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.722994] env[62952]: DEBUG nova.network.neutron [req-62637716-10c8-404a-a986-cc58e5a1c33b req-c34f670c-8b93-47d2-abb1-e68d111de230 service nova] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Refreshing network info cache for port d4ed51c5-2a91-40ed-8ce9-62f6109ab3b8 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 884.731838] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e62a7e-4cb0-1995-9826-de9c8290e280, 'name': SearchDatastore_Task, 'duration_secs': 0.044717} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.732953] env[62952]: DEBUG oslo_concurrency.lockutils [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.733389] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 884.733726] env[62952]: DEBUG oslo_concurrency.lockutils [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.733936] env[62952]: DEBUG oslo_concurrency.lockutils [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.734180] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 884.734705] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-651c38b1-6fec-4729-a06e-687d3f3907a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.747093] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 884.747450] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 884.748443] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce35be7f-3f92-497c-a8ce-84d4f8efe14b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.758156] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 884.758156] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524b67a8-e14e-c10a-383a-dce5f95a524d" [ 884.758156] env[62952]: _type = "Task" [ 884.758156] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.766434] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263256, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517205} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.767129] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 6e289c20-8dd2-4680-b816-9bbe82fe7aec/6e289c20-8dd2-4680-b816-9bbe82fe7aec.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 884.767440] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 884.768100] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-188da270-db33-4847-9c13-bf4efaad29a6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.774557] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524b67a8-e14e-c10a-383a-dce5f95a524d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.780449] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 884.780449] env[62952]: value = "task-1263258" [ 884.780449] env[62952]: _type = "Task" [ 884.780449] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.790262] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263258, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.970754] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Acquiring lock "8b594aff-3c11-4cc1-8c47-9c317f32a12d" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.970754] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Lock "8b594aff-3c11-4cc1-8c47-9c317f32a12d" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.097503] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "refresh_cache-3416eb38-b961-4d48-8b37-8b7b92375dc9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.114834] env[62952]: DEBUG oslo_concurrency.lockutils [req-e69be5bc-2b8e-421d-8c25-5eefaf3261dd req-174c1a15-77eb-4079-a360-7e7fcaef1ae8 service nova] Releasing lock "refresh_cache-dd700ee9-0957-4210-be50-e4c8ac3ca456" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.146862] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Releasing lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.166386] env[62952]: INFO nova.compute.manager [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Took 36.02 seconds to build instance. [ 885.257269] env[62952]: DEBUG nova.network.neutron [req-62637716-10c8-404a-a986-cc58e5a1c33b req-c34f670c-8b93-47d2-abb1-e68d111de230 service nova] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 885.271187] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524b67a8-e14e-c10a-383a-dce5f95a524d, 'name': SearchDatastore_Task, 'duration_secs': 0.016004} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.272091] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72502054-1074-4590-879c-add6477e0cea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.278399] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 885.278399] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c23335-bad0-1eaa-ac30-17984a69e6b6" [ 885.278399] env[62952]: _type = "Task" [ 885.278399] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.294851] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c23335-bad0-1eaa-ac30-17984a69e6b6, 'name': SearchDatastore_Task, 'duration_secs': 0.010369} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.295164] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263258, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.261764} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.295420] env[62952]: DEBUG oslo_concurrency.lockutils [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.295673] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] dd700ee9-0957-4210-be50-e4c8ac3ca456/dd700ee9-0957-4210-be50-e4c8ac3ca456.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 885.296138] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 885.296234] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b72f446-e2e9-43fd-87c6-b34177a1aa5d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.298589] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71942dba-7992-44e3-a4d4-a1188ace887d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.323969] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 6e289c20-8dd2-4680-b816-9bbe82fe7aec/6e289c20-8dd2-4680-b816-9bbe82fe7aec.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 885.328292] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c68d233d-50e8-4013-9174-f94c43aa3bf7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.343463] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 885.343463] env[62952]: value = "task-1263259" [ 885.343463] env[62952]: _type = "Task" [ 885.343463] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.350992] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 885.350992] env[62952]: value = "task-1263260" [ 885.350992] env[62952]: _type = "Task" [ 885.350992] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.354219] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263259, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.362909] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263260, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.384907] env[62952]: DEBUG nova.network.neutron [req-62637716-10c8-404a-a986-cc58e5a1c33b req-c34f670c-8b93-47d2-abb1-e68d111de230 service nova] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.476288] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Lock "8b594aff-3c11-4cc1-8c47-9c317f32a12d" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.505s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.476904] env[62952]: DEBUG nova.compute.manager [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 885.668923] env[62952]: DEBUG oslo_concurrency.lockutils [None req-746821d5-8ebb-48c6-9199-c7e6e5baf34d tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Lock "2b11ad9b-63d8-40da-8f68-3d8260c63bc7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.170s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.679611] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 885.680098] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2260decb-8b40-4cf9-bcde-376096b72cbe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.689112] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 885.689112] env[62952]: value = "task-1263261" [ 885.689112] env[62952]: _type = "Task" [ 885.689112] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.704133] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263261, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.857634] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263259, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454022} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.860922] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] dd700ee9-0957-4210-be50-e4c8ac3ca456/dd700ee9-0957-4210-be50-e4c8ac3ca456.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 885.861168] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 885.862398] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-74831966-ac44-484f-94ae-748aee4ad211 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.865057] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b00076d3-139f-48ce-b580-fd35b3687479 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.877458] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7595914e-e531-4349-91ee-f3a162ef9d87 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.881083] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263260, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.882722] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 885.882722] env[62952]: value = "task-1263262" [ 885.882722] env[62952]: _type = "Task" [ 885.882722] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.887307] env[62952]: DEBUG oslo_concurrency.lockutils [req-62637716-10c8-404a-a986-cc58e5a1c33b req-c34f670c-8b93-47d2-abb1-e68d111de230 service nova] Releasing lock "refresh_cache-3416eb38-b961-4d48-8b37-8b7b92375dc9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.888064] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "refresh_cache-3416eb38-b961-4d48-8b37-8b7b92375dc9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.888278] env[62952]: DEBUG nova.network.neutron [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 885.925537] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb47d00f-c108-4f43-b74c-2af65a4fe5dd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.928641] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263262, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.935682] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aabb7be-548b-4ea4-ba22-940cf969e2b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.953504] env[62952]: DEBUG nova.compute.provider_tree [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.960019] env[62952]: DEBUG nova.network.neutron [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 885.985702] env[62952]: DEBUG nova.compute.utils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.987582] env[62952]: DEBUG nova.compute.manager [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 885.987582] env[62952]: DEBUG nova.network.neutron [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 886.056785] env[62952]: DEBUG nova.policy [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e200cfcca41c4c808b282cb984a5c5d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9a219c7158e444b1a56227c5b5afebd7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 886.170321] env[62952]: DEBUG nova.network.neutron [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Updating instance_info_cache with network_info: [{"id": "d4ed51c5-2a91-40ed-8ce9-62f6109ab3b8", "address": "fa:16:3e:9c:97:d6", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4ed51c5-2a", "ovs_interfaceid": "d4ed51c5-2a91-40ed-8ce9-62f6109ab3b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.171804] env[62952]: DEBUG nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 886.199802] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263261, 'name': PowerOffVM_Task, 'duration_secs': 0.26154} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.200048] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 886.200866] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd1a57a-5dc5-481f-97da-993a5ef6cdbb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.220584] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b29084-42e8-4abf-8fc9-ed3991f7b8d4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.251536] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 886.251823] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-529487cd-762a-4f0e-989b-fd65e9224b17 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.260261] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 886.260261] env[62952]: value = "task-1263263" [ 886.260261] env[62952]: _type = "Task" [ 886.260261] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.268476] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263263, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.359323] env[62952]: DEBUG nova.network.neutron [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Successfully created port: 968b1d86-0e93-4f9a-9003-392cfa822d2c {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.372253] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263260, 'name': ReconfigVM_Task, 'duration_secs': 0.564487} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.376976] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 6e289c20-8dd2-4680-b816-9bbe82fe7aec/6e289c20-8dd2-4680-b816-9bbe82fe7aec.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.376976] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9d4f0f3e-d1b2-452f-a286-124ebda58aa1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.386218] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 886.386218] env[62952]: value = "task-1263264" [ 886.386218] env[62952]: _type = "Task" [ 886.386218] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.397763] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263262, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077089} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.400896] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 886.401261] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263264, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.401979] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e292df1-5249-4678-9a49-aa72a4a53972 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.426621] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] dd700ee9-0957-4210-be50-e4c8ac3ca456/dd700ee9-0957-4210-be50-e4c8ac3ca456.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.426621] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae80b420-9a96-4300-a09c-24a0f7d3fa50 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.449180] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 886.449180] env[62952]: value = "task-1263265" [ 886.449180] env[62952]: _type = "Task" [ 886.449180] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.458976] env[62952]: DEBUG nova.scheduler.client.report [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.462462] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263265, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.490849] env[62952]: DEBUG nova.compute.manager [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 886.528173] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 886.528434] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271870', 'volume_id': '42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d', 'name': 'volume-42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f4f8bee4-bf92-4321-afdc-16f7fb17dbae', 'attached_at': '', 'detached_at': '', 'volume_id': '42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d', 'serial': '42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 886.529467] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c866aa6f-527a-4456-909a-e0dc0c10b9d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.548820] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01aa53b-ba46-47c3-950e-1c3651da35ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.576114] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] volume-42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d/volume-42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.576481] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d5f12c0-3780-47df-94fb-79d050548bbd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.599055] env[62952]: DEBUG oslo_vmware.api [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 886.599055] env[62952]: value = "task-1263266" [ 886.599055] env[62952]: _type = "Task" [ 886.599055] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.610594] env[62952]: DEBUG oslo_vmware.api [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263266, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.676270] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "refresh_cache-3416eb38-b961-4d48-8b37-8b7b92375dc9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.676739] env[62952]: DEBUG nova.compute.manager [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Instance network_info: |[{"id": "d4ed51c5-2a91-40ed-8ce9-62f6109ab3b8", "address": "fa:16:3e:9c:97:d6", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4ed51c5-2a", "ovs_interfaceid": "d4ed51c5-2a91-40ed-8ce9-62f6109ab3b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 886.679581] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:97:d6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5ad8894f-e240-4013-8272-4e79daea0751', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd4ed51c5-2a91-40ed-8ce9-62f6109ab3b8', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 886.687185] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Creating folder: Project (fbb5ccb4d615429facd6932fb55b04e8). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 886.689730] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-075acad6-3a86-4da1-8091-d70e89b9e928 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.702034] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Created folder: Project (fbb5ccb4d615429facd6932fb55b04e8) in parent group-v271811. [ 886.702318] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Creating folder: Instances. Parent ref: group-v271874. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 886.703244] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.703490] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6be84542-a694-42c8-91db-555366eeddd1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.716488] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Created folder: Instances in parent group-v271874. [ 886.716750] env[62952]: DEBUG oslo.service.loopingcall [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.716977] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 886.717217] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-921c38b3-31e2-443c-b56a-0c6f539c08ed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.738609] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 886.738609] env[62952]: value = "task-1263269" [ 886.738609] env[62952]: _type = "Task" [ 886.738609] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.748218] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263269, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.774293] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 886.774452] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 886.774650] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.774808] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.775055] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 886.775388] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-61edecf3-0eb7-4578-83b6-72c88d8a5ab4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.795810] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 886.795810] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 886.796211] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-184f55d7-b8c0-41b7-ae11-023a129a2f30 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.802702] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 886.802702] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52922ee5-3e73-d696-b617-ef36133f4fb5" [ 886.802702] env[62952]: _type = "Task" [ 886.802702] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.812929] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52922ee5-3e73-d696-b617-ef36133f4fb5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.902036] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263264, 'name': Rename_Task, 'duration_secs': 0.202031} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.902336] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 886.902600] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-14584f8e-6070-485a-ac14-21ab03717f13 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.910859] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 886.910859] env[62952]: value = "task-1263270" [ 886.910859] env[62952]: _type = "Task" [ 886.910859] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.922588] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263270, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.961878] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263265, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.965202] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.500s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.966102] env[62952]: DEBUG nova.compute.manager [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 886.970137] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.975s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.970353] env[62952]: DEBUG nova.objects.instance [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62952) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 887.111387] env[62952]: DEBUG oslo_vmware.api [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263266, 'name': ReconfigVM_Task, 'duration_secs': 0.451436} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.111781] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Reconfigured VM instance instance-00000031 to attach disk [datastore1] volume-42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d/volume-42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.117329] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66d2989d-f9c4-4f10-910b-b60a397e4d7e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.135956] env[62952]: DEBUG oslo_vmware.api [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 887.135956] env[62952]: value = "task-1263271" [ 887.135956] env[62952]: _type = "Task" [ 887.135956] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.145021] env[62952]: DEBUG oslo_vmware.api [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263271, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.248561] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263269, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.314669] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52922ee5-3e73-d696-b617-ef36133f4fb5, 'name': SearchDatastore_Task, 'duration_secs': 0.015278} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.315150] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69f3d116-b61a-4273-b683-59370d74e334 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.320692] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 887.320692] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52178a38-5193-2205-b120-c5246815ecea" [ 887.320692] env[62952]: _type = "Task" [ 887.320692] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.329269] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52178a38-5193-2205-b120-c5246815ecea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.420878] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263270, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.460270] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263265, 'name': ReconfigVM_Task, 'duration_secs': 0.563755} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.460597] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Reconfigured VM instance instance-0000003d to attach disk [datastore1] dd700ee9-0957-4210-be50-e4c8ac3ca456/dd700ee9-0957-4210-be50-e4c8ac3ca456.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.461251] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7dd821e3-9245-47dd-942b-5d522366cfd4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.469255] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 887.469255] env[62952]: value = "task-1263272" [ 887.469255] env[62952]: _type = "Task" [ 887.469255] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.473344] env[62952]: DEBUG nova.compute.utils [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 887.474793] env[62952]: DEBUG nova.compute.manager [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 887.475183] env[62952]: DEBUG nova.network.neutron [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 887.487605] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263272, 'name': Rename_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.502340] env[62952]: DEBUG nova.compute.manager [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 887.523488] env[62952]: DEBUG nova.policy [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '469eae9db3d14e9f90988e3093514e69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff928791a3cf496cb8db2c2a6c47a6a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 887.528611] env[62952]: DEBUG nova.virt.hardware [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.528859] env[62952]: DEBUG nova.virt.hardware [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.529043] env[62952]: DEBUG nova.virt.hardware [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.529253] env[62952]: DEBUG nova.virt.hardware [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.529405] env[62952]: DEBUG nova.virt.hardware [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.529557] env[62952]: DEBUG nova.virt.hardware [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.529770] env[62952]: DEBUG nova.virt.hardware [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.529934] env[62952]: DEBUG nova.virt.hardware [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.530124] env[62952]: DEBUG nova.virt.hardware [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.530295] env[62952]: DEBUG nova.virt.hardware [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.530472] env[62952]: DEBUG nova.virt.hardware [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.531383] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94542698-6636-49ab-a501-6609eceb4c62 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.540235] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7407423b-981e-42ca-9ee3-f57e210139c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.648679] env[62952]: DEBUG oslo_vmware.api [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263271, 'name': ReconfigVM_Task, 'duration_secs': 0.257747} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.649088] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271870', 'volume_id': '42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d', 'name': 'volume-42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f4f8bee4-bf92-4321-afdc-16f7fb17dbae', 'attached_at': '', 'detached_at': '', 'volume_id': '42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d', 'serial': '42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 887.750630] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263269, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.832786] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52178a38-5193-2205-b120-c5246815ecea, 'name': SearchDatastore_Task, 'duration_secs': 0.012716} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.833022] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.833251] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 67b729be-fb66-4ed1-bbea-e62216d460d5/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk. {{(pid=62952) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 887.833525] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5d09e0d9-44e7-4619-9ccb-458dda8622b5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.842362] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 887.842362] env[62952]: value = "task-1263273" [ 887.842362] env[62952]: _type = "Task" [ 887.842362] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.853255] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263273, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.922638] env[62952]: DEBUG oslo_vmware.api [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263270, 'name': PowerOnVM_Task, 'duration_secs': 0.758066} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.923697] env[62952]: DEBUG nova.network.neutron [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Successfully created port: ac76c133-49eb-4a8a-afe8-015a43840974 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 887.926027] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 887.926449] env[62952]: INFO nova.compute.manager [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Took 8.88 seconds to spawn the instance on the hypervisor. [ 887.926677] env[62952]: DEBUG nova.compute.manager [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 887.927691] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a651eef2-9bc7-4304-acb8-d911bd7c3aee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.950422] env[62952]: DEBUG nova.compute.manager [req-aef9626c-ef24-4893-99f4-4628a55aba53 req-0e65fad4-d92d-4a4c-aeb0-ec87eae1b01a service nova] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Received event network-vif-plugged-968b1d86-0e93-4f9a-9003-392cfa822d2c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 887.950708] env[62952]: DEBUG oslo_concurrency.lockutils [req-aef9626c-ef24-4893-99f4-4628a55aba53 req-0e65fad4-d92d-4a4c-aeb0-ec87eae1b01a service nova] Acquiring lock "49947bde-069b-459c-ae19-ae5f090f535b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.950908] env[62952]: DEBUG oslo_concurrency.lockutils [req-aef9626c-ef24-4893-99f4-4628a55aba53 req-0e65fad4-d92d-4a4c-aeb0-ec87eae1b01a service nova] Lock "49947bde-069b-459c-ae19-ae5f090f535b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.951100] env[62952]: DEBUG oslo_concurrency.lockutils [req-aef9626c-ef24-4893-99f4-4628a55aba53 req-0e65fad4-d92d-4a4c-aeb0-ec87eae1b01a service nova] Lock "49947bde-069b-459c-ae19-ae5f090f535b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.951277] env[62952]: DEBUG nova.compute.manager [req-aef9626c-ef24-4893-99f4-4628a55aba53 req-0e65fad4-d92d-4a4c-aeb0-ec87eae1b01a service nova] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] No waiting events found dispatching network-vif-plugged-968b1d86-0e93-4f9a-9003-392cfa822d2c {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 887.951445] env[62952]: WARNING nova.compute.manager [req-aef9626c-ef24-4893-99f4-4628a55aba53 req-0e65fad4-d92d-4a4c-aeb0-ec87eae1b01a service nova] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Received unexpected event network-vif-plugged-968b1d86-0e93-4f9a-9003-392cfa822d2c for instance with vm_state building and task_state spawning. [ 887.981395] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263272, 'name': Rename_Task, 'duration_secs': 0.323226} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.982615] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ce82f1a-2e9a-41d4-b29a-61aeb9835427 tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.984346] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 887.985299] env[62952]: DEBUG nova.compute.manager [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 887.987733] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.868s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.989593] env[62952]: INFO nova.compute.claims [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 887.992408] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e571e72-fe7d-4dee-97a8-f8719f0976c4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.004614] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 888.004614] env[62952]: value = "task-1263274" [ 888.004614] env[62952]: _type = "Task" [ 888.004614] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.014725] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263274, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.154724] env[62952]: DEBUG nova.network.neutron [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Successfully updated port: 968b1d86-0e93-4f9a-9003-392cfa822d2c {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 888.170039] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquiring lock "2b11ad9b-63d8-40da-8f68-3d8260c63bc7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.170241] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Lock "2b11ad9b-63d8-40da-8f68-3d8260c63bc7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.170339] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquiring lock "2b11ad9b-63d8-40da-8f68-3d8260c63bc7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.170666] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Lock "2b11ad9b-63d8-40da-8f68-3d8260c63bc7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.170666] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Lock "2b11ad9b-63d8-40da-8f68-3d8260c63bc7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.176914] env[62952]: INFO nova.compute.manager [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Terminating instance [ 888.179555] env[62952]: DEBUG nova.compute.manager [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 888.181310] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 888.182678] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d25205-cfa2-4bf7-9331-06d34e741c93 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.200630] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 888.200965] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c741ab83-3d4d-4d90-8db0-db2b7736b17a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.209719] env[62952]: DEBUG oslo_vmware.api [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Waiting for the task: (returnval){ [ 888.209719] env[62952]: value = "task-1263275" [ 888.209719] env[62952]: _type = "Task" [ 888.209719] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.221681] env[62952]: DEBUG oslo_vmware.api [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': task-1263275, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.253944] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263269, 'name': CreateVM_Task, 'duration_secs': 1.387773} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.254568] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 888.255221] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.255398] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.255760] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 888.256072] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-446e7436-829d-448e-8861-2423fd76d391 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.262316] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 888.262316] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5288ac3a-5f42-f5ba-bcaf-523843054b98" [ 888.262316] env[62952]: _type = "Task" [ 888.262316] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.272079] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5288ac3a-5f42-f5ba-bcaf-523843054b98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.353907] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263273, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.449254] env[62952]: INFO nova.compute.manager [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Took 37.27 seconds to build instance. [ 888.515950] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263274, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.665861] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Acquiring lock "refresh_cache-49947bde-069b-459c-ae19-ae5f090f535b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.665861] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Acquired lock "refresh_cache-49947bde-069b-459c-ae19-ae5f090f535b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.665861] env[62952]: DEBUG nova.network.neutron [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 888.722473] env[62952]: DEBUG oslo_vmware.api [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': task-1263275, 'name': PowerOffVM_Task, 'duration_secs': 0.222522} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.723051] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 888.723051] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 888.723406] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-105e9640-95d3-4b99-86b6-823be5c31297 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.726954] env[62952]: DEBUG nova.objects.instance [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lazy-loading 'flavor' on Instance uuid f4f8bee4-bf92-4321-afdc-16f7fb17dbae {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.772866] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5288ac3a-5f42-f5ba-bcaf-523843054b98, 'name': SearchDatastore_Task, 'duration_secs': 0.063219} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.773552] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.773552] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.773937] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.773937] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.774064] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.774686] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7a2ae60c-089e-41f7-91b8-dcd434209c89 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.785699] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.786110] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 888.786675] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b5b0617-7b3c-4721-b157-c4a8ce85a0d9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.794495] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 888.794495] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52cae774-3500-8d3d-0d51-0e5bcab12b44" [ 888.794495] env[62952]: _type = "Task" [ 888.794495] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.809131] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52cae774-3500-8d3d-0d51-0e5bcab12b44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.810598] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 888.810802] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 888.811326] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Deleting the datastore file [datastore1] 2b11ad9b-63d8-40da-8f68-3d8260c63bc7 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 888.811326] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5707f32b-2ee6-4369-9da3-7bb8f886574b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.821667] env[62952]: DEBUG oslo_vmware.api [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Waiting for the task: (returnval){ [ 888.821667] env[62952]: value = "task-1263277" [ 888.821667] env[62952]: _type = "Task" [ 888.821667] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.830513] env[62952]: DEBUG oslo_vmware.api [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': task-1263277, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.862991] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263273, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.673885} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.863279] env[62952]: INFO nova.virt.vmwareapi.ds_util [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 67b729be-fb66-4ed1-bbea-e62216d460d5/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk. [ 888.864842] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f66b605-e37e-44a9-82cc-6f88d48f9e27 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.893906] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 67b729be-fb66-4ed1-bbea-e62216d460d5/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.894046] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0051c034-bed5-46f7-ac09-489846cf9d1c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.916699] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 888.916699] env[62952]: value = "task-1263278" [ 888.916699] env[62952]: _type = "Task" [ 888.916699] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.926519] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263278, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.950954] env[62952]: DEBUG oslo_concurrency.lockutils [None req-514f6688-6b15-46d2-b376-c881db15ce9c tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.340s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.004879] env[62952]: DEBUG nova.compute.manager [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 889.017802] env[62952]: DEBUG oslo_vmware.api [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263274, 'name': PowerOnVM_Task, 'duration_secs': 0.611903} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.019611] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 889.019611] env[62952]: INFO nova.compute.manager [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Took 7.47 seconds to spawn the instance on the hypervisor. [ 889.019611] env[62952]: DEBUG nova.compute.manager [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 889.019787] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e091ed-9d55-4d3a-a09b-dbd2595d4b18 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.039853] env[62952]: DEBUG nova.virt.hardware [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 889.040127] env[62952]: DEBUG nova.virt.hardware [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 889.040325] env[62952]: DEBUG nova.virt.hardware [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 889.040540] env[62952]: DEBUG nova.virt.hardware [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 889.040687] env[62952]: DEBUG nova.virt.hardware [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 889.040866] env[62952]: DEBUG nova.virt.hardware [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 889.041068] env[62952]: DEBUG nova.virt.hardware [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 889.041240] env[62952]: DEBUG nova.virt.hardware [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 889.041521] env[62952]: DEBUG nova.virt.hardware [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 889.041774] env[62952]: DEBUG nova.virt.hardware [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 889.041987] env[62952]: DEBUG nova.virt.hardware [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 889.042879] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4098bd93-67cf-48eb-ad73-6283fd77041a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.054963] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a30528d-f773-4c4f-978c-1b0a58dc909b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.215281] env[62952]: DEBUG nova.network.neutron [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 889.234212] env[62952]: DEBUG oslo_concurrency.lockutils [None req-127e8119-af94-4a57-9fe1-5da4ec376c0e tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.320s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.308256] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52cae774-3500-8d3d-0d51-0e5bcab12b44, 'name': SearchDatastore_Task, 'duration_secs': 0.016429} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.309040] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87213fde-8b8b-4e5d-99bb-237bea1e14ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.317155] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 889.317155] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c1e942-5eae-7f74-f5c6-79eeab2e02c6" [ 889.317155] env[62952]: _type = "Task" [ 889.317155] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.325185] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c1e942-5eae-7f74-f5c6-79eeab2e02c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.337660] env[62952]: DEBUG oslo_concurrency.lockutils [None req-263deb02-7fc8-4ce6-ab8e-0dad53b320ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.337660] env[62952]: DEBUG oslo_concurrency.lockutils [None req-263deb02-7fc8-4ce6-ab8e-0dad53b320ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.337660] env[62952]: DEBUG nova.compute.manager [None req-263deb02-7fc8-4ce6-ab8e-0dad53b320ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 889.337660] env[62952]: DEBUG oslo_vmware.api [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Task: {'id': task-1263277, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.281922} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.337660] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0fcff43-4e8c-4dd7-a6d8-3a381bc0ee2a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.339552] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 889.339750] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 889.339916] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 889.340304] env[62952]: INFO nova.compute.manager [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Took 1.16 seconds to destroy the instance on the hypervisor. [ 889.340371] env[62952]: DEBUG oslo.service.loopingcall [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.340704] env[62952]: DEBUG nova.compute.manager [-] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 889.340704] env[62952]: DEBUG nova.network.neutron [-] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 889.346677] env[62952]: DEBUG nova.compute.manager [None req-263deb02-7fc8-4ce6-ab8e-0dad53b320ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62952) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 889.346677] env[62952]: DEBUG nova.objects.instance [None req-263deb02-7fc8-4ce6-ab8e-0dad53b320ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lazy-loading 'flavor' on Instance uuid f4f8bee4-bf92-4321-afdc-16f7fb17dbae {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.384493] env[62952]: DEBUG nova.network.neutron [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Updating instance_info_cache with network_info: [{"id": "968b1d86-0e93-4f9a-9003-392cfa822d2c", "address": "fa:16:3e:a3:20:14", "network": {"id": "0f7b7452-68eb-4a36-a028-090a06854581", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-45238576-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a219c7158e444b1a56227c5b5afebd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89ed4797-90ad-44cd-bbcb-e90b2a8400f3", "external-id": "nsx-vlan-transportzone-699", "segmentation_id": 699, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap968b1d86-0e", "ovs_interfaceid": "968b1d86-0e93-4f9a-9003-392cfa822d2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.414109] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69395c69-f1e5-4160-8277-776500547bee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.423824] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4da9e82-1249-4bd4-8ba3-a1df143e9688 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.430032] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263278, 'name': ReconfigVM_Task, 'duration_secs': 0.352714} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.430686] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 67b729be-fb66-4ed1-bbea-e62216d460d5/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 889.431521] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9930c73-da08-4853-8f36-c48ee6728654 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.461617] env[62952]: DEBUG nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 889.465694] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139a0c45-92b5-4b8e-b0b3-54d2570499dc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.491873] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-19419aa3-cd7d-4bfb-a24f-5d4897d725a9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.506487] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42eb7ee1-b75a-4818-b9d7-7ff86cdda1cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.516552] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 889.516552] env[62952]: value = "task-1263279" [ 889.516552] env[62952]: _type = "Task" [ 889.516552] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.528421] env[62952]: DEBUG nova.compute.provider_tree [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 889.543394] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263279, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.543394] env[62952]: INFO nova.compute.manager [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Took 37.71 seconds to build instance. [ 889.837175] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c1e942-5eae-7f74-f5c6-79eeab2e02c6, 'name': SearchDatastore_Task, 'duration_secs': 0.057533} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.838475] env[62952]: DEBUG nova.network.neutron [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Successfully updated port: ac76c133-49eb-4a8a-afe8-015a43840974 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 889.840687] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.841529] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 3416eb38-b961-4d48-8b37-8b7b92375dc9/3416eb38-b961-4d48-8b37-8b7b92375dc9.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 889.842479] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f933e255-dee4-4033-93a7-415cf6680b6d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.854379] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-263deb02-7fc8-4ce6-ab8e-0dad53b320ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 889.855760] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b1b6ea35-9fbd-450b-a6c5-0b3014a5f2ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.860749] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 889.860749] env[62952]: value = "task-1263280" [ 889.860749] env[62952]: _type = "Task" [ 889.860749] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.868412] env[62952]: DEBUG oslo_vmware.api [None req-263deb02-7fc8-4ce6-ab8e-0dad53b320ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 889.868412] env[62952]: value = "task-1263281" [ 889.868412] env[62952]: _type = "Task" [ 889.868412] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.877875] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263280, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.884582] env[62952]: DEBUG oslo_vmware.api [None req-263deb02-7fc8-4ce6-ab8e-0dad53b320ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263281, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.886203] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Releasing lock "refresh_cache-49947bde-069b-459c-ae19-ae5f090f535b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.886525] env[62952]: DEBUG nova.compute.manager [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Instance network_info: |[{"id": "968b1d86-0e93-4f9a-9003-392cfa822d2c", "address": "fa:16:3e:a3:20:14", "network": {"id": "0f7b7452-68eb-4a36-a028-090a06854581", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-45238576-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a219c7158e444b1a56227c5b5afebd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89ed4797-90ad-44cd-bbcb-e90b2a8400f3", "external-id": "nsx-vlan-transportzone-699", "segmentation_id": 699, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap968b1d86-0e", "ovs_interfaceid": "968b1d86-0e93-4f9a-9003-392cfa822d2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 889.888246] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:20:14', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '89ed4797-90ad-44cd-bbcb-e90b2a8400f3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '968b1d86-0e93-4f9a-9003-392cfa822d2c', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 889.896030] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Creating folder: Project (9a219c7158e444b1a56227c5b5afebd7). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 889.896697] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-658d8d05-7b5d-4bcc-b163-a6f21aa7a808 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.908452] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Created folder: Project (9a219c7158e444b1a56227c5b5afebd7) in parent group-v271811. [ 889.908660] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Creating folder: Instances. Parent ref: group-v271877. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 889.908990] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1a172c41-a05d-4720-948d-d8e7f0a3d1a7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.918991] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Created folder: Instances in parent group-v271877. [ 889.919391] env[62952]: DEBUG oslo.service.loopingcall [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.919693] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 889.920952] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f49c69de-bdef-4670-87ea-dd70eb95a259 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.942483] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 889.942483] env[62952]: value = "task-1263284" [ 889.942483] env[62952]: _type = "Task" [ 889.942483] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.951978] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263284, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.987443] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.994019] env[62952]: DEBUG nova.compute.manager [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Received event network-changed-968b1d86-0e93-4f9a-9003-392cfa822d2c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 889.994257] env[62952]: DEBUG nova.compute.manager [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Refreshing instance network info cache due to event network-changed-968b1d86-0e93-4f9a-9003-392cfa822d2c. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 889.994487] env[62952]: DEBUG oslo_concurrency.lockutils [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] Acquiring lock "refresh_cache-49947bde-069b-459c-ae19-ae5f090f535b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.994945] env[62952]: DEBUG oslo_concurrency.lockutils [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] Acquired lock "refresh_cache-49947bde-069b-459c-ae19-ae5f090f535b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.994945] env[62952]: DEBUG nova.network.neutron [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Refreshing network info cache for port 968b1d86-0e93-4f9a-9003-392cfa822d2c {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 890.027555] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263279, 'name': ReconfigVM_Task, 'duration_secs': 0.163125} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.027836] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 890.028113] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3b8d9855-d5de-436b-93c5-549a9419e182 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.031144] env[62952]: DEBUG nova.scheduler.client.report [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.036117] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 890.036117] env[62952]: value = "task-1263285" [ 890.036117] env[62952]: _type = "Task" [ 890.036117] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.047800] env[62952]: DEBUG oslo_concurrency.lockutils [None req-125ae2dc-143c-4ef6-8b2f-a6138b38ae11 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "dd700ee9-0957-4210-be50-e4c8ac3ca456" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.435s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.048101] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263285, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.312517] env[62952]: DEBUG nova.network.neutron [-] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.341695] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "refresh_cache-ba022bea-e357-4541-a573-d1da2eb49b46" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.341924] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "refresh_cache-ba022bea-e357-4541-a573-d1da2eb49b46" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.342302] env[62952]: DEBUG nova.network.neutron [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 890.355542] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "dd700ee9-0957-4210-be50-e4c8ac3ca456" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.355800] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "dd700ee9-0957-4210-be50-e4c8ac3ca456" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.355982] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "dd700ee9-0957-4210-be50-e4c8ac3ca456-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.356185] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "dd700ee9-0957-4210-be50-e4c8ac3ca456-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.356354] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "dd700ee9-0957-4210-be50-e4c8ac3ca456-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.358557] env[62952]: INFO nova.compute.manager [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Terminating instance [ 890.361193] env[62952]: DEBUG nova.compute.manager [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 890.361396] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 890.362387] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-861e68ed-96e7-4b15-bcc4-8ad72c7b5f35 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.374877] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 890.378478] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d14cfc55-157a-4e77-a23b-b72646228385 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.380390] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263280, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.385025] env[62952]: DEBUG oslo_vmware.api [None req-263deb02-7fc8-4ce6-ab8e-0dad53b320ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263281, 'name': PowerOffVM_Task, 'duration_secs': 0.407204} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.385025] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-263deb02-7fc8-4ce6-ab8e-0dad53b320ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 890.385025] env[62952]: DEBUG nova.compute.manager [None req-263deb02-7fc8-4ce6-ab8e-0dad53b320ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 890.385811] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6156755c-43da-423b-bce3-3a0a95f9d235 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.389947] env[62952]: DEBUG oslo_vmware.api [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 890.389947] env[62952]: value = "task-1263286" [ 890.389947] env[62952]: _type = "Task" [ 890.389947] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.409319] env[62952]: DEBUG oslo_vmware.api [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263286, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.458946] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263284, 'name': CreateVM_Task, 'duration_secs': 0.410173} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.458946] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 890.459562] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.459778] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.460216] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 890.460563] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9dda2998-7b2e-417e-b81b-d824c989ef96 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.469727] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Waiting for the task: (returnval){ [ 890.469727] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52318fee-af71-ef1b-509e-96b017594f9b" [ 890.469727] env[62952]: _type = "Task" [ 890.469727] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.480478] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52318fee-af71-ef1b-509e-96b017594f9b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.535697] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.548s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.536045] env[62952]: DEBUG nova.compute.manager [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 890.539414] env[62952]: DEBUG oslo_concurrency.lockutils [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.247s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.540580] env[62952]: INFO nova.compute.claims [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 890.551570] env[62952]: DEBUG nova.compute.manager [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 890.559597] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263285, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.751369] env[62952]: DEBUG nova.network.neutron [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Updated VIF entry in instance network info cache for port 968b1d86-0e93-4f9a-9003-392cfa822d2c. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 890.751834] env[62952]: DEBUG nova.network.neutron [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Updating instance_info_cache with network_info: [{"id": "968b1d86-0e93-4f9a-9003-392cfa822d2c", "address": "fa:16:3e:a3:20:14", "network": {"id": "0f7b7452-68eb-4a36-a028-090a06854581", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-45238576-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a219c7158e444b1a56227c5b5afebd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89ed4797-90ad-44cd-bbcb-e90b2a8400f3", "external-id": "nsx-vlan-transportzone-699", "segmentation_id": 699, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap968b1d86-0e", "ovs_interfaceid": "968b1d86-0e93-4f9a-9003-392cfa822d2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.815964] env[62952]: INFO nova.compute.manager [-] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Took 1.48 seconds to deallocate network for instance. [ 890.875722] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263280, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.756443} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.875722] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 3416eb38-b961-4d48-8b37-8b7b92375dc9/3416eb38-b961-4d48-8b37-8b7b92375dc9.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 890.875722] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 890.875722] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-27fc97b1-8895-49c6-83b5-e6e1b4f77a62 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.877907] env[62952]: DEBUG nova.network.neutron [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 890.886130] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 890.886130] env[62952]: value = "task-1263287" [ 890.886130] env[62952]: _type = "Task" [ 890.886130] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.894523] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263287, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.907949] env[62952]: DEBUG oslo_vmware.api [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263286, 'name': PowerOffVM_Task, 'duration_secs': 0.239745} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.908451] env[62952]: DEBUG oslo_concurrency.lockutils [None req-263deb02-7fc8-4ce6-ab8e-0dad53b320ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.573s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.909416] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 890.909611] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 890.910075] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8532d508-666a-4eb3-9b7b-a2ce4d287e36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.982171] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52318fee-af71-ef1b-509e-96b017594f9b, 'name': SearchDatastore_Task, 'duration_secs': 0.02882} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.983180] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.983940] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.983940] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.983940] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.984117] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.984268] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 890.984449] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 890.984894] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleting the datastore file [datastore1] dd700ee9-0957-4210-be50-e4c8ac3ca456 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 890.984894] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b9dd5469-826d-47b4-b4aa-58d73ca0078f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.986958] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ce0f6c00-2b9c-4954-9d49-840e2ba43808 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.995611] env[62952]: DEBUG oslo_vmware.api [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 890.995611] env[62952]: value = "task-1263289" [ 890.995611] env[62952]: _type = "Task" [ 890.995611] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.004548] env[62952]: DEBUG oslo_vmware.api [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263289, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.006659] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.006858] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 891.009710] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca3ea078-50b9-4fad-b358-32541c751d67 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.015726] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Waiting for the task: (returnval){ [ 891.015726] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520d9b14-e413-a4d7-212b-b39f79382d31" [ 891.015726] env[62952]: _type = "Task" [ 891.015726] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.024181] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520d9b14-e413-a4d7-212b-b39f79382d31, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.044939] env[62952]: DEBUG nova.compute.utils [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 891.050178] env[62952]: DEBUG nova.compute.manager [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 891.050178] env[62952]: DEBUG nova.network.neutron [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 891.056763] env[62952]: DEBUG nova.network.neutron [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Updating instance_info_cache with network_info: [{"id": "ac76c133-49eb-4a8a-afe8-015a43840974", "address": "fa:16:3e:8e:01:a3", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac76c133-49", "ovs_interfaceid": "ac76c133-49eb-4a8a-afe8-015a43840974", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.063116] env[62952]: DEBUG oslo_vmware.api [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263285, 'name': PowerOnVM_Task, 'duration_secs': 0.651757} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.066362] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 891.068800] env[62952]: DEBUG nova.compute.manager [None req-6d17876f-2ccd-43e7-ac46-b3d4db32c90b tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 891.069863] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b6b3f0-38f1-4f2c-8ad6-f5c934996d45 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.083881] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.117097] env[62952]: DEBUG nova.policy [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6b8f4e5921074d81b584ec6860a714dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '59b5f1ca72394f58a12e0fb9942615c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 891.237839] env[62952]: DEBUG oslo_concurrency.lockutils [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.237839] env[62952]: DEBUG oslo_concurrency.lockutils [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.237839] env[62952]: INFO nova.compute.manager [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Shelving [ 891.254081] env[62952]: DEBUG oslo_concurrency.lockutils [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] Releasing lock "refresh_cache-49947bde-069b-459c-ae19-ae5f090f535b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.254797] env[62952]: DEBUG nova.compute.manager [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Received event network-changed-5e2e9d1e-dce4-404e-bf41-1a7897be3758 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 891.255041] env[62952]: DEBUG nova.compute.manager [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Refreshing instance network info cache due to event network-changed-5e2e9d1e-dce4-404e-bf41-1a7897be3758. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 891.255308] env[62952]: DEBUG oslo_concurrency.lockutils [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] Acquiring lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.255492] env[62952]: DEBUG oslo_concurrency.lockutils [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] Acquired lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.255691] env[62952]: DEBUG nova.network.neutron [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Refreshing network info cache for port 5e2e9d1e-dce4-404e-bf41-1a7897be3758 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 891.326358] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.396763] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263287, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.128601} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.397081] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 891.398111] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d7a7c8-f84f-4a54-8e87-0760a02d713b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.428825] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 3416eb38-b961-4d48-8b37-8b7b92375dc9/3416eb38-b961-4d48-8b37-8b7b92375dc9.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 891.429260] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57e4ca9f-19e9-49e5-adeb-bf485c8059fb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.457348] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 891.457348] env[62952]: value = "task-1263290" [ 891.457348] env[62952]: _type = "Task" [ 891.457348] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.465214] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263290, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.507152] env[62952]: DEBUG oslo_vmware.api [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263289, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.531703] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520d9b14-e413-a4d7-212b-b39f79382d31, 'name': SearchDatastore_Task, 'duration_secs': 0.017953} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.532539] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f91fac8-a4f1-4f91-bfc8-e03db5d32472 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.539524] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Waiting for the task: (returnval){ [ 891.539524] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5244f27a-0c56-6e88-c2fc-94823c1ac2da" [ 891.539524] env[62952]: _type = "Task" [ 891.539524] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.550318] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5244f27a-0c56-6e88-c2fc-94823c1ac2da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.553157] env[62952]: DEBUG nova.compute.manager [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 891.567190] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "refresh_cache-ba022bea-e357-4541-a573-d1da2eb49b46" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.567469] env[62952]: DEBUG nova.compute.manager [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Instance network_info: |[{"id": "ac76c133-49eb-4a8a-afe8-015a43840974", "address": "fa:16:3e:8e:01:a3", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac76c133-49", "ovs_interfaceid": "ac76c133-49eb-4a8a-afe8-015a43840974", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 891.567878] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:01:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ac76c133-49eb-4a8a-afe8-015a43840974', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 891.575817] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Creating folder: Project (ff928791a3cf496cb8db2c2a6c47a6a6). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 891.576743] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f2cc40e-6adf-4b30-bf7c-4069d90571ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.593068] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Created folder: Project (ff928791a3cf496cb8db2c2a6c47a6a6) in parent group-v271811. [ 891.593068] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Creating folder: Instances. Parent ref: group-v271880. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 891.593068] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-34f5e1c3-7a38-4e0c-b73e-d9f8b557d7f1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.607032] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Created folder: Instances in parent group-v271880. [ 891.607719] env[62952]: DEBUG oslo.service.loopingcall [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 891.608823] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 891.610962] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-24b2a737-b31a-45ab-bdb1-4405da1e03e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.630367] env[62952]: DEBUG nova.network.neutron [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Successfully created port: f9a55a9d-b041-4065-b354-8a180d765835 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 891.641740] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 891.641740] env[62952]: value = "task-1263293" [ 891.641740] env[62952]: _type = "Task" [ 891.641740] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.651354] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263293, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.746223] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 891.746513] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b188d5ee-261b-4ae9-963d-d8b4a613022e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.761331] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 891.761331] env[62952]: value = "task-1263294" [ 891.761331] env[62952]: _type = "Task" [ 891.761331] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.773572] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263294, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.969379] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263290, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.002094] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b73bca8-7164-4084-8ec1-8fc4d84959ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.010727] env[62952]: DEBUG oslo_vmware.api [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263289, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.70699} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.010897] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 892.011107] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 892.011355] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 892.011566] env[62952]: INFO nova.compute.manager [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Took 1.65 seconds to destroy the instance on the hypervisor. [ 892.011853] env[62952]: DEBUG oslo.service.loopingcall [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.012079] env[62952]: DEBUG nova.compute.manager [-] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 892.012155] env[62952]: DEBUG nova.network.neutron [-] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 892.018085] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b451128-6ee3-4317-9da5-ac30adb85005 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.055273] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-772da4d1-ef3f-433f-8153-9df1e0532f5b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.057409] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Acquiring lock "219cfb14-ca73-45d3-86e8-ed4642d7d480" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.058049] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Lock "219cfb14-ca73-45d3-86e8-ed4642d7d480" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.058049] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Acquiring lock "219cfb14-ca73-45d3-86e8-ed4642d7d480-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.058049] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Lock "219cfb14-ca73-45d3-86e8-ed4642d7d480-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.058212] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Lock "219cfb14-ca73-45d3-86e8-ed4642d7d480-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.063536] env[62952]: INFO nova.compute.manager [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Terminating instance [ 892.066171] env[62952]: DEBUG nova.compute.manager [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 892.066390] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 892.067236] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e13466-ccb9-41c9-8639-ea042d4cd4e8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.075985] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5244f27a-0c56-6e88-c2fc-94823c1ac2da, 'name': SearchDatastore_Task, 'duration_secs': 0.033742} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.076769] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.076903] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 49947bde-069b-459c-ae19-ae5f090f535b/49947bde-069b-459c-ae19-ae5f090f535b.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 892.078135] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4848787f-22d4-44d4-8441-c97c52875a7c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.084075] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f0ee046-e6be-48a6-8381-65035d13031b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.086306] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 892.086890] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-202401d1-18c0-46d7-b9cd-8105805462ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.098261] env[62952]: DEBUG nova.compute.provider_tree [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.103422] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Waiting for the task: (returnval){ [ 892.103422] env[62952]: value = "task-1263295" [ 892.103422] env[62952]: _type = "Task" [ 892.103422] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.103683] env[62952]: DEBUG oslo_vmware.api [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Waiting for the task: (returnval){ [ 892.103683] env[62952]: value = "task-1263296" [ 892.103683] env[62952]: _type = "Task" [ 892.103683] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.118057] env[62952]: DEBUG oslo_vmware.api [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': task-1263296, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.121389] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': task-1263295, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.133809] env[62952]: DEBUG nova.objects.instance [None req-8c0eee7c-3ce3-40b3-b6ba-183b3bf25f2c tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lazy-loading 'flavor' on Instance uuid f4f8bee4-bf92-4321-afdc-16f7fb17dbae {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.154155] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263293, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.188398] env[62952]: DEBUG nova.network.neutron [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updated VIF entry in instance network info cache for port 5e2e9d1e-dce4-404e-bf41-1a7897be3758. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 892.188823] env[62952]: DEBUG nova.network.neutron [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance_info_cache with network_info: [{"id": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "address": "fa:16:3e:eb:c2:8d", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e2e9d1e-dc", "ovs_interfaceid": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.274664] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263294, 'name': PowerOffVM_Task, 'duration_secs': 0.212662} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.274794] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 892.276064] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9451cc91-8fef-4b8f-a6b7-822e2d49fe04 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.306599] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eff0298-8c9a-48bb-9467-f5ef96838681 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.411923] env[62952]: DEBUG nova.compute.manager [req-fd4a0e6a-b296-4dc0-9e85-a0fc8f0d4581 req-b3873b43-5db1-49e1-8350-98197d133432 service nova] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Received event network-vif-deleted-dfcafefe-49c2-461d-b13c-9186be103ed0 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 892.412155] env[62952]: INFO nova.compute.manager [req-fd4a0e6a-b296-4dc0-9e85-a0fc8f0d4581 req-b3873b43-5db1-49e1-8350-98197d133432 service nova] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Neutron deleted interface dfcafefe-49c2-461d-b13c-9186be103ed0; detaching it from the instance and deleting it from the info cache [ 892.412337] env[62952]: DEBUG nova.network.neutron [req-fd4a0e6a-b296-4dc0-9e85-a0fc8f0d4581 req-b3873b43-5db1-49e1-8350-98197d133432 service nova] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.468945] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263290, 'name': ReconfigVM_Task, 'duration_secs': 0.529023} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.472278] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 3416eb38-b961-4d48-8b37-8b7b92375dc9/3416eb38-b961-4d48-8b37-8b7b92375dc9.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 892.472278] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cc2cc29f-16f1-44d1-b224-98cf88fa755f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.482031] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 892.482031] env[62952]: value = "task-1263297" [ 892.482031] env[62952]: _type = "Task" [ 892.482031] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.494583] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263297, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.568786] env[62952]: DEBUG nova.compute.manager [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 892.601463] env[62952]: DEBUG nova.virt.hardware [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 892.601979] env[62952]: DEBUG nova.virt.hardware [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 892.602320] env[62952]: DEBUG nova.virt.hardware [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 892.602674] env[62952]: DEBUG nova.virt.hardware [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 892.603024] env[62952]: DEBUG nova.virt.hardware [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 892.606086] env[62952]: DEBUG nova.virt.hardware [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 892.606086] env[62952]: DEBUG nova.virt.hardware [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 892.606086] env[62952]: DEBUG nova.virt.hardware [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 892.606086] env[62952]: DEBUG nova.virt.hardware [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 892.606086] env[62952]: DEBUG nova.virt.hardware [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 892.606256] env[62952]: DEBUG nova.virt.hardware [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 892.606256] env[62952]: DEBUG nova.scheduler.client.report [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 892.609342] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e09b8d1-e943-4bca-9ad2-b756d80d89bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.627893] env[62952]: DEBUG oslo_vmware.api [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': task-1263296, 'name': PowerOffVM_Task, 'duration_secs': 0.263871} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.634959] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 892.635944] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 892.635944] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': task-1263295, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.637059] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22b38cfd-efd8-472e-b761-171e93294be8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.640064] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84cf445-82e2-453c-bf2a-8845ed045169 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.651325] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c0eee7c-3ce3-40b3-b6ba-183b3bf25f2c tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.651325] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c0eee7c-3ce3-40b3-b6ba-183b3bf25f2c tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquired lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.651325] env[62952]: DEBUG nova.network.neutron [None req-8c0eee7c-3ce3-40b3-b6ba-183b3bf25f2c tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 892.651325] env[62952]: DEBUG nova.objects.instance [None req-8c0eee7c-3ce3-40b3-b6ba-183b3bf25f2c tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lazy-loading 'info_cache' on Instance uuid f4f8bee4-bf92-4321-afdc-16f7fb17dbae {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.669680] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263293, 'name': CreateVM_Task, 'duration_secs': 0.619562} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.669878] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 892.670666] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.670925] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.671357] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 892.671704] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-528ce704-96e7-468c-8a04-6d386050a9ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.679582] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 892.679582] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c90082-4934-558d-268d-07b8023e586f" [ 892.679582] env[62952]: _type = "Task" [ 892.679582] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.692022] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c90082-4934-558d-268d-07b8023e586f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.692621] env[62952]: DEBUG oslo_concurrency.lockutils [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] Releasing lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.692950] env[62952]: DEBUG nova.compute.manager [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Received event network-vif-plugged-ac76c133-49eb-4a8a-afe8-015a43840974 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 892.693216] env[62952]: DEBUG oslo_concurrency.lockutils [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] Acquiring lock "ba022bea-e357-4541-a573-d1da2eb49b46-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.693722] env[62952]: DEBUG oslo_concurrency.lockutils [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] Lock "ba022bea-e357-4541-a573-d1da2eb49b46-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.694316] env[62952]: DEBUG oslo_concurrency.lockutils [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] Lock "ba022bea-e357-4541-a573-d1da2eb49b46-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.694316] env[62952]: DEBUG nova.compute.manager [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] No waiting events found dispatching network-vif-plugged-ac76c133-49eb-4a8a-afe8-015a43840974 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 892.694446] env[62952]: WARNING nova.compute.manager [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Received unexpected event network-vif-plugged-ac76c133-49eb-4a8a-afe8-015a43840974 for instance with vm_state building and task_state spawning. [ 892.694589] env[62952]: DEBUG nova.compute.manager [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Received event network-vif-deleted-484ad815-dff6-400c-93dd-babc8b8b5c9c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 892.694798] env[62952]: INFO nova.compute.manager [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Neutron deleted interface 484ad815-dff6-400c-93dd-babc8b8b5c9c; detaching it from the instance and deleting it from the info cache [ 892.695057] env[62952]: DEBUG nova.network.neutron [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.740419] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 892.740924] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 892.741230] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Deleting the datastore file [datastore1] 219cfb14-ca73-45d3-86e8-ed4642d7d480 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 892.741706] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-06c26fe5-fbbd-4961-8872-bb4691b7a568 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.753975] env[62952]: DEBUG oslo_vmware.api [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Waiting for the task: (returnval){ [ 892.753975] env[62952]: value = "task-1263299" [ 892.753975] env[62952]: _type = "Task" [ 892.753975] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.771432] env[62952]: DEBUG oslo_vmware.api [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': task-1263299, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.821194] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 892.821535] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4c1f83ba-88fe-46b7-a8f1-83524029dec1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.831632] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 892.831632] env[62952]: value = "task-1263300" [ 892.831632] env[62952]: _type = "Task" [ 892.831632] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.837526] env[62952]: DEBUG nova.network.neutron [-] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.842168] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263300, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.915483] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d7473919-4a3c-40c9-8467-3afcb6b49480 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.926676] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0434a0f7-1932-488f-9fd8-41428c4076c1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.963817] env[62952]: DEBUG nova.compute.manager [req-fd4a0e6a-b296-4dc0-9e85-a0fc8f0d4581 req-b3873b43-5db1-49e1-8350-98197d133432 service nova] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Detach interface failed, port_id=dfcafefe-49c2-461d-b13c-9186be103ed0, reason: Instance dd700ee9-0957-4210-be50-e4c8ac3ca456 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 892.992876] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263297, 'name': Rename_Task, 'duration_secs': 0.190324} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.995772] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 892.995772] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ca13362-ce40-4996-8875-fced6904cb0b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.001163] env[62952]: DEBUG nova.compute.manager [req-8fe1c27f-891d-49b2-a65b-698a2ff62dff req-42162fc5-2364-43e1-8ea7-43f48f1b4df9 service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Received event network-changed-3423511f-8828-4fca-aec2-79e921a53579 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 893.001163] env[62952]: DEBUG nova.compute.manager [req-8fe1c27f-891d-49b2-a65b-698a2ff62dff req-42162fc5-2364-43e1-8ea7-43f48f1b4df9 service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Refreshing instance network info cache due to event network-changed-3423511f-8828-4fca-aec2-79e921a53579. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 893.001163] env[62952]: DEBUG oslo_concurrency.lockutils [req-8fe1c27f-891d-49b2-a65b-698a2ff62dff req-42162fc5-2364-43e1-8ea7-43f48f1b4df9 service nova] Acquiring lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.001163] env[62952]: DEBUG oslo_concurrency.lockutils [req-8fe1c27f-891d-49b2-a65b-698a2ff62dff req-42162fc5-2364-43e1-8ea7-43f48f1b4df9 service nova] Acquired lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.001163] env[62952]: DEBUG nova.network.neutron [req-8fe1c27f-891d-49b2-a65b-698a2ff62dff req-42162fc5-2364-43e1-8ea7-43f48f1b4df9 service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Refreshing network info cache for port 3423511f-8828-4fca-aec2-79e921a53579 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 893.008189] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 893.008189] env[62952]: value = "task-1263301" [ 893.008189] env[62952]: _type = "Task" [ 893.008189] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.018811] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263301, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.114443] env[62952]: DEBUG oslo_concurrency.lockutils [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.575s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.114928] env[62952]: DEBUG nova.compute.manager [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 893.118602] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.066s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.118842] env[62952]: DEBUG nova.objects.instance [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Lazy-loading 'resources' on Instance uuid bc369b94-2f40-444e-88d7-069fd33e2646 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.133704] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': task-1263295, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.702821} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.134186] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 49947bde-069b-459c-ae19-ae5f090f535b/49947bde-069b-459c-ae19-ae5f090f535b.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 893.134760] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.136569] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b219e319-935a-421d-b749-94c415155cd6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.149505] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Waiting for the task: (returnval){ [ 893.149505] env[62952]: value = "task-1263302" [ 893.149505] env[62952]: _type = "Task" [ 893.149505] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.160484] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': task-1263302, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.166198] env[62952]: DEBUG nova.objects.base [None req-8c0eee7c-3ce3-40b3-b6ba-183b3bf25f2c tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 893.194340] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c90082-4934-558d-268d-07b8023e586f, 'name': SearchDatastore_Task, 'duration_secs': 0.064307} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.194884] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.195213] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 893.195468] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.195621] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.195998] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 893.196300] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5f1d0dac-c508-4c20-b85d-bc4713685627 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.199142] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-64f86317-ca4f-47d4-9d27-6c620c041d77 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.210027] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af15f7d-7741-49c1-938c-8784b3af6855 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.225080] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 893.225296] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 893.226456] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c247a6d-ff56-47d4-847e-64c653a998ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.232211] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 893.232211] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f41a90-0ee5-4442-7849-0fa3d70bcb50" [ 893.232211] env[62952]: _type = "Task" [ 893.232211] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.248181] env[62952]: DEBUG nova.compute.manager [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Detach interface failed, port_id=484ad815-dff6-400c-93dd-babc8b8b5c9c, reason: Instance 2b11ad9b-63d8-40da-8f68-3d8260c63bc7 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 893.248346] env[62952]: DEBUG nova.compute.manager [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Received event network-changed-ac76c133-49eb-4a8a-afe8-015a43840974 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 893.248538] env[62952]: DEBUG nova.compute.manager [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Refreshing instance network info cache due to event network-changed-ac76c133-49eb-4a8a-afe8-015a43840974. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 893.248770] env[62952]: DEBUG oslo_concurrency.lockutils [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] Acquiring lock "refresh_cache-ba022bea-e357-4541-a573-d1da2eb49b46" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.248919] env[62952]: DEBUG oslo_concurrency.lockutils [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] Acquired lock "refresh_cache-ba022bea-e357-4541-a573-d1da2eb49b46" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.249350] env[62952]: DEBUG nova.network.neutron [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Refreshing network info cache for port ac76c133-49eb-4a8a-afe8-015a43840974 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 893.254864] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f41a90-0ee5-4442-7849-0fa3d70bcb50, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.264961] env[62952]: DEBUG oslo_vmware.api [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Task: {'id': task-1263299, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176494} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.265238] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 893.265422] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 893.265601] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 893.265795] env[62952]: INFO nova.compute.manager [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Took 1.20 seconds to destroy the instance on the hypervisor. [ 893.266055] env[62952]: DEBUG oslo.service.loopingcall [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 893.266253] env[62952]: DEBUG nova.compute.manager [-] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 893.266347] env[62952]: DEBUG nova.network.neutron [-] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 893.326669] env[62952]: DEBUG nova.network.neutron [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Successfully updated port: f9a55a9d-b041-4065-b354-8a180d765835 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 893.344437] env[62952]: INFO nova.compute.manager [-] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Took 1.33 seconds to deallocate network for instance. [ 893.344798] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263300, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.522119] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263301, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.626859] env[62952]: DEBUG nova.compute.utils [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 893.633359] env[62952]: DEBUG nova.compute.manager [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 893.633359] env[62952]: DEBUG nova.network.neutron [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 893.666819] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': task-1263302, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075098} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.670588] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.677122] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e39038d4-ed1a-4f6c-9c69-2ab0e770014c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.716370] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 49947bde-069b-459c-ae19-ae5f090f535b/49947bde-069b-459c-ae19-ae5f090f535b.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.716986] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6263e90-4122-4229-937e-10e23fcc441d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.746521] env[62952]: DEBUG nova.policy [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6b8f4e5921074d81b584ec6860a714dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '59b5f1ca72394f58a12e0fb9942615c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 893.752163] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Waiting for the task: (returnval){ [ 893.752163] env[62952]: value = "task-1263303" [ 893.752163] env[62952]: _type = "Task" [ 893.752163] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.761018] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f41a90-0ee5-4442-7849-0fa3d70bcb50, 'name': SearchDatastore_Task, 'duration_secs': 0.013238} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.762384] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-401b6a6f-783e-4ea8-91e1-6cad24637407 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.771550] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': task-1263303, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.775756] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 893.775756] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52aad78e-3c20-0b5d-0a5c-0eb77279d5b6" [ 893.775756] env[62952]: _type = "Task" [ 893.775756] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.784867] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52aad78e-3c20-0b5d-0a5c-0eb77279d5b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.827631] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquiring lock "refresh_cache-14f8dbbe-2c4a-4948-bf15-106d5b1b0677" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.827631] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquired lock "refresh_cache-14f8dbbe-2c4a-4948-bf15-106d5b1b0677" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.827631] env[62952]: DEBUG nova.network.neutron [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 893.843822] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263300, 'name': CreateSnapshot_Task, 'duration_secs': 0.880737} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.846571] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 893.850056] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a511bbb-e23b-47d1-a6b0-edb3e6518da4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.854217] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.019214] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263301, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.037065] env[62952]: DEBUG nova.network.neutron [req-8fe1c27f-891d-49b2-a65b-698a2ff62dff req-42162fc5-2364-43e1-8ea7-43f48f1b4df9 service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Updated VIF entry in instance network info cache for port 3423511f-8828-4fca-aec2-79e921a53579. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 894.037512] env[62952]: DEBUG nova.network.neutron [req-8fe1c27f-891d-49b2-a65b-698a2ff62dff req-42162fc5-2364-43e1-8ea7-43f48f1b4df9 service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Updating instance_info_cache with network_info: [{"id": "3423511f-8828-4fca-aec2-79e921a53579", "address": "fa:16:3e:3b:bd:29", "network": {"id": "d865ccd7-b8a3-49d7-811f-6d4466bbbf45", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1549996301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4484b1cb78ff4e7baac4b9a5d1e63549", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a5569c57-a27b-4d4d-a519-784be105114f", "external-id": "nsx-vlan-transportzone-640", "segmentation_id": 640, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3423511f-88", "ovs_interfaceid": "3423511f-8828-4fca-aec2-79e921a53579", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.075233] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426e1bfa-1876-405c-822b-c8df7b2025a2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.083689] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262feded-a991-4996-8f2a-fedaaea4dd96 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.117779] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b7458a-472f-4382-81b0-d8ad286d05fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.127604] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad6b434-594f-4ea0-9be2-a118e3bba5b4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.132907] env[62952]: DEBUG nova.compute.manager [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 894.144425] env[62952]: DEBUG nova.compute.provider_tree [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.215799] env[62952]: DEBUG nova.network.neutron [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Successfully created port: 31694ebf-fc67-4662-98f6-9d99b17eb865 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 894.263062] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': task-1263303, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.287251] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52aad78e-3c20-0b5d-0a5c-0eb77279d5b6, 'name': SearchDatastore_Task, 'duration_secs': 0.016503} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.287625] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.287865] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] ba022bea-e357-4541-a573-d1da2eb49b46/ba022bea-e357-4541-a573-d1da2eb49b46.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 894.288152] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ebc6582c-d65c-4986-9c29-d43df6cfecc8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.297417] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 894.297417] env[62952]: value = "task-1263304" [ 894.297417] env[62952]: _type = "Task" [ 894.297417] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.307233] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263304, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.370795] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 894.371320] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ce6cb582-6d4a-4e68-98de-5d53adcc7931 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.376092] env[62952]: DEBUG nova.network.neutron [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 894.381486] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 894.381486] env[62952]: value = "task-1263305" [ 894.381486] env[62952]: _type = "Task" [ 894.381486] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.390362] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263305, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.523100] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263301, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.541736] env[62952]: DEBUG oslo_concurrency.lockutils [req-8fe1c27f-891d-49b2-a65b-698a2ff62dff req-42162fc5-2364-43e1-8ea7-43f48f1b4df9 service nova] Releasing lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.589574] env[62952]: DEBUG nova.network.neutron [None req-8c0eee7c-3ce3-40b3-b6ba-183b3bf25f2c tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Updating instance_info_cache with network_info: [{"id": "363ca83f-c6ac-4446-996c-edc459eb29bf", "address": "fa:16:3e:dd:39:7e", "network": {"id": "3c430a3e-4ea2-4c7b-8e05-733f4a306f04", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-441856160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8960909bb80f4395bb87b6de3c60ab2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap363ca83f-c6", "ovs_interfaceid": "363ca83f-c6ac-4446-996c-edc459eb29bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.599167] env[62952]: DEBUG nova.network.neutron [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Updating instance_info_cache with network_info: [{"id": "f9a55a9d-b041-4065-b354-8a180d765835", "address": "fa:16:3e:37:37:40", "network": {"id": "2e11ee50-4303-4f0a-a9fc-464f5f2087b3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "73931ed9ad374fb0870629249c5f6909", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3c995e9-7f2f-420c-880a-d60da6e708ad", "external-id": "nsx-vlan-transportzone-166", "segmentation_id": 166, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9a55a9d-b0", "ovs_interfaceid": "f9a55a9d-b041-4065-b354-8a180d765835", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.617857] env[62952]: DEBUG nova.network.neutron [-] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.637932] env[62952]: DEBUG nova.network.neutron [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Updated VIF entry in instance network info cache for port ac76c133-49eb-4a8a-afe8-015a43840974. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 894.638255] env[62952]: DEBUG nova.network.neutron [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Updating instance_info_cache with network_info: [{"id": "ac76c133-49eb-4a8a-afe8-015a43840974", "address": "fa:16:3e:8e:01:a3", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac76c133-49", "ovs_interfaceid": "ac76c133-49eb-4a8a-afe8-015a43840974", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.654798] env[62952]: DEBUG nova.scheduler.client.report [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 894.763413] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': task-1263303, 'name': ReconfigVM_Task, 'duration_secs': 0.590965} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.763736] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 49947bde-069b-459c-ae19-ae5f090f535b/49947bde-069b-459c-ae19-ae5f090f535b.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.764544] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bb1f4b73-50eb-4a02-8f9f-74d4dd6e78f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.773505] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Waiting for the task: (returnval){ [ 894.773505] env[62952]: value = "task-1263306" [ 894.773505] env[62952]: _type = "Task" [ 894.773505] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.784295] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': task-1263306, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.810110] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263304, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.895146] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263305, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.022348] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263301, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.092632] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c0eee7c-3ce3-40b3-b6ba-183b3bf25f2c tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Releasing lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.101920] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Releasing lock "refresh_cache-14f8dbbe-2c4a-4948-bf15-106d5b1b0677" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.105172] env[62952]: DEBUG nova.compute.manager [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Instance network_info: |[{"id": "f9a55a9d-b041-4065-b354-8a180d765835", "address": "fa:16:3e:37:37:40", "network": {"id": "2e11ee50-4303-4f0a-a9fc-464f5f2087b3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "73931ed9ad374fb0870629249c5f6909", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3c995e9-7f2f-420c-880a-d60da6e708ad", "external-id": "nsx-vlan-transportzone-166", "segmentation_id": 166, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9a55a9d-b0", "ovs_interfaceid": "f9a55a9d-b041-4065-b354-8a180d765835", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 895.105391] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:37:37:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3c995e9-7f2f-420c-880a-d60da6e708ad', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f9a55a9d-b041-4065-b354-8a180d765835', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 895.112688] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Creating folder: Project (59b5f1ca72394f58a12e0fb9942615c5). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 895.113888] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3bbe9f0c-b963-4fd4-8276-bb01347495de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.120627] env[62952]: INFO nova.compute.manager [-] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Took 1.85 seconds to deallocate network for instance. [ 895.127653] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Created folder: Project (59b5f1ca72394f58a12e0fb9942615c5) in parent group-v271811. [ 895.127858] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Creating folder: Instances. Parent ref: group-v271885. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 895.128349] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2f1e4912-4f71-4873-99a0-d695a634411c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.134553] env[62952]: DEBUG nova.compute.manager [req-00264ebc-52c3-4fbb-9c5e-35d29d56a5bd req-b08b2c41-b22b-4f29-ae3a-3b621fbc2288 service nova] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Received event network-vif-plugged-f9a55a9d-b041-4065-b354-8a180d765835 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 895.134553] env[62952]: DEBUG oslo_concurrency.lockutils [req-00264ebc-52c3-4fbb-9c5e-35d29d56a5bd req-b08b2c41-b22b-4f29-ae3a-3b621fbc2288 service nova] Acquiring lock "14f8dbbe-2c4a-4948-bf15-106d5b1b0677-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.134553] env[62952]: DEBUG oslo_concurrency.lockutils [req-00264ebc-52c3-4fbb-9c5e-35d29d56a5bd req-b08b2c41-b22b-4f29-ae3a-3b621fbc2288 service nova] Lock "14f8dbbe-2c4a-4948-bf15-106d5b1b0677-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.134553] env[62952]: DEBUG oslo_concurrency.lockutils [req-00264ebc-52c3-4fbb-9c5e-35d29d56a5bd req-b08b2c41-b22b-4f29-ae3a-3b621fbc2288 service nova] Lock "14f8dbbe-2c4a-4948-bf15-106d5b1b0677-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.134553] env[62952]: DEBUG nova.compute.manager [req-00264ebc-52c3-4fbb-9c5e-35d29d56a5bd req-b08b2c41-b22b-4f29-ae3a-3b621fbc2288 service nova] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] No waiting events found dispatching network-vif-plugged-f9a55a9d-b041-4065-b354-8a180d765835 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 895.134826] env[62952]: WARNING nova.compute.manager [req-00264ebc-52c3-4fbb-9c5e-35d29d56a5bd req-b08b2c41-b22b-4f29-ae3a-3b621fbc2288 service nova] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Received unexpected event network-vif-plugged-f9a55a9d-b041-4065-b354-8a180d765835 for instance with vm_state building and task_state spawning. [ 895.134826] env[62952]: DEBUG nova.compute.manager [req-00264ebc-52c3-4fbb-9c5e-35d29d56a5bd req-b08b2c41-b22b-4f29-ae3a-3b621fbc2288 service nova] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Received event network-changed-f9a55a9d-b041-4065-b354-8a180d765835 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 895.134826] env[62952]: DEBUG nova.compute.manager [req-00264ebc-52c3-4fbb-9c5e-35d29d56a5bd req-b08b2c41-b22b-4f29-ae3a-3b621fbc2288 service nova] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Refreshing instance network info cache due to event network-changed-f9a55a9d-b041-4065-b354-8a180d765835. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 895.134928] env[62952]: DEBUG oslo_concurrency.lockutils [req-00264ebc-52c3-4fbb-9c5e-35d29d56a5bd req-b08b2c41-b22b-4f29-ae3a-3b621fbc2288 service nova] Acquiring lock "refresh_cache-14f8dbbe-2c4a-4948-bf15-106d5b1b0677" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.134995] env[62952]: DEBUG oslo_concurrency.lockutils [req-00264ebc-52c3-4fbb-9c5e-35d29d56a5bd req-b08b2c41-b22b-4f29-ae3a-3b621fbc2288 service nova] Acquired lock "refresh_cache-14f8dbbe-2c4a-4948-bf15-106d5b1b0677" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.135940] env[62952]: DEBUG nova.network.neutron [req-00264ebc-52c3-4fbb-9c5e-35d29d56a5bd req-b08b2c41-b22b-4f29-ae3a-3b621fbc2288 service nova] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Refreshing network info cache for port f9a55a9d-b041-4065-b354-8a180d765835 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 895.140610] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Created folder: Instances in parent group-v271885. [ 895.140870] env[62952]: DEBUG oslo.service.loopingcall [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.142185] env[62952]: DEBUG oslo_concurrency.lockutils [req-cc5bff6a-e5c1-4951-b0b0-02487dff810d req-cd32dcaf-cb07-498e-bb08-da2b2d2e68f9 service nova] Releasing lock "refresh_cache-ba022bea-e357-4541-a573-d1da2eb49b46" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.142185] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 895.142185] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-855c341a-d02d-4f9b-8a35-f12da7a56ba4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.160039] env[62952]: DEBUG nova.compute.manager [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 895.167378] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.049s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.171995] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.481s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.176025] env[62952]: INFO nova.compute.claims [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 895.176025] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 895.176025] env[62952]: value = "task-1263309" [ 895.176025] env[62952]: _type = "Task" [ 895.176025] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.188728] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263309, 'name': CreateVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.191365] env[62952]: DEBUG nova.virt.hardware [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 895.191512] env[62952]: DEBUG nova.virt.hardware [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 895.191617] env[62952]: DEBUG nova.virt.hardware [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 895.191824] env[62952]: DEBUG nova.virt.hardware [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 895.191966] env[62952]: DEBUG nova.virt.hardware [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 895.192133] env[62952]: DEBUG nova.virt.hardware [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 895.192363] env[62952]: DEBUG nova.virt.hardware [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 895.192605] env[62952]: DEBUG nova.virt.hardware [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 895.193453] env[62952]: DEBUG nova.virt.hardware [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 895.193453] env[62952]: DEBUG nova.virt.hardware [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 895.193453] env[62952]: DEBUG nova.virt.hardware [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 895.194180] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7226b18-1472-4736-8837-d05f65327bba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.200564] env[62952]: INFO nova.scheduler.client.report [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Deleted allocations for instance bc369b94-2f40-444e-88d7-069fd33e2646 [ 895.210672] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a690982-0b68-4008-84e2-91ed5b90ecf0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.287206] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': task-1263306, 'name': Rename_Task, 'duration_secs': 0.331769} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.287592] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 895.288050] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4a805e67-12a4-4d1f-bef4-cb066823ac6c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.295812] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Waiting for the task: (returnval){ [ 895.295812] env[62952]: value = "task-1263310" [ 895.295812] env[62952]: _type = "Task" [ 895.295812] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.307784] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': task-1263310, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.310929] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263304, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.637751} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.311235] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] ba022bea-e357-4541-a573-d1da2eb49b46/ba022bea-e357-4541-a573-d1da2eb49b46.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 895.311491] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 895.311769] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ddbb0212-a81e-431b-9ef7-374a8214fae3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.318944] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 895.318944] env[62952]: value = "task-1263311" [ 895.318944] env[62952]: _type = "Task" [ 895.318944] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.327959] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263311, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.393235] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263305, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.521774] env[62952]: DEBUG oslo_vmware.api [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263301, 'name': PowerOnVM_Task, 'duration_secs': 2.01849} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.523169] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 895.523452] env[62952]: INFO nova.compute.manager [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Took 11.52 seconds to spawn the instance on the hypervisor. [ 895.523784] env[62952]: DEBUG nova.compute.manager [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.525056] env[62952]: DEBUG nova.compute.manager [req-3ed14ab1-355c-4bca-bad5-799387d10b61 req-60f5b9ec-3bb9-4fe5-98ee-1e08aa4efb4f service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Received event network-changed-3423511f-8828-4fca-aec2-79e921a53579 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 895.525306] env[62952]: DEBUG nova.compute.manager [req-3ed14ab1-355c-4bca-bad5-799387d10b61 req-60f5b9ec-3bb9-4fe5-98ee-1e08aa4efb4f service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Refreshing instance network info cache due to event network-changed-3423511f-8828-4fca-aec2-79e921a53579. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 895.525635] env[62952]: DEBUG oslo_concurrency.lockutils [req-3ed14ab1-355c-4bca-bad5-799387d10b61 req-60f5b9ec-3bb9-4fe5-98ee-1e08aa4efb4f service nova] Acquiring lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.525771] env[62952]: DEBUG oslo_concurrency.lockutils [req-3ed14ab1-355c-4bca-bad5-799387d10b61 req-60f5b9ec-3bb9-4fe5-98ee-1e08aa4efb4f service nova] Acquired lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.525948] env[62952]: DEBUG nova.network.neutron [req-3ed14ab1-355c-4bca-bad5-799387d10b61 req-60f5b9ec-3bb9-4fe5-98ee-1e08aa4efb4f service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Refreshing network info cache for port 3423511f-8828-4fca-aec2-79e921a53579 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 895.527854] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab937a31-7b15-40c0-9d32-a518ae747fb8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.597092] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c0eee7c-3ce3-40b3-b6ba-183b3bf25f2c tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 895.597641] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aadf76c2-06f9-416d-9b73-366551aa9b71 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.606749] env[62952]: DEBUG oslo_vmware.api [None req-8c0eee7c-3ce3-40b3-b6ba-183b3bf25f2c tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 895.606749] env[62952]: value = "task-1263312" [ 895.606749] env[62952]: _type = "Task" [ 895.606749] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.616544] env[62952]: DEBUG oslo_vmware.api [None req-8c0eee7c-3ce3-40b3-b6ba-183b3bf25f2c tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263312, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.629924] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.689995] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263309, 'name': CreateVM_Task, 'duration_secs': 0.440931} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.690204] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 895.691021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.691112] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.691535] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 895.692351] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-524fc07e-68e4-4516-99a7-a67528184036 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.698416] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 895.698416] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c0800b-8acd-09ba-0617-f8eaaa09092c" [ 895.698416] env[62952]: _type = "Task" [ 895.698416] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.707928] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c0800b-8acd-09ba-0617-f8eaaa09092c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.721229] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9679bf57-a721-4a3b-9820-ec6faa9b76da tempest-ServerShowV257Test-192688974 tempest-ServerShowV257Test-192688974-project-member] Lock "bc369b94-2f40-444e-88d7-069fd33e2646" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.378s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.809561] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': task-1263310, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.831844] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263311, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107504} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.832196] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 895.833074] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd517c2-6048-487c-b410-26b9d84baaef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.863482] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] ba022bea-e357-4541-a573-d1da2eb49b46/ba022bea-e357-4541-a573-d1da2eb49b46.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 895.866300] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13606c62-83b2-4489-b88b-1d8175b0d9d6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.885857] env[62952]: DEBUG nova.network.neutron [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Successfully updated port: 31694ebf-fc67-4662-98f6-9d99b17eb865 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 895.901563] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263305, 'name': CloneVM_Task, 'duration_secs': 1.49085} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.901770] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 895.901770] env[62952]: value = "task-1263313" [ 895.901770] env[62952]: _type = "Task" [ 895.901770] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.902016] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Created linked-clone VM from snapshot [ 895.902814] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee42b0be-900c-42d8-9c6f-c9b70c433761 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.916574] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Uploading image 315e9b97-1fa4-4f98-9bd7-cd708e6717cd {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 895.921927] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263313, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.943935] env[62952]: DEBUG oslo_vmware.rw_handles [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 895.943935] env[62952]: value = "vm-271884" [ 895.943935] env[62952]: _type = "VirtualMachine" [ 895.943935] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 895.944302] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-195e73bb-28fe-4ba3-9ea6-f9cecca7ba5b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.953793] env[62952]: DEBUG oslo_vmware.rw_handles [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lease: (returnval){ [ 895.953793] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b14733-2d84-8e59-186e-446a9d949651" [ 895.953793] env[62952]: _type = "HttpNfcLease" [ 895.953793] env[62952]: } obtained for exporting VM: (result){ [ 895.953793] env[62952]: value = "vm-271884" [ 895.953793] env[62952]: _type = "VirtualMachine" [ 895.953793] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 895.954183] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the lease: (returnval){ [ 895.954183] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b14733-2d84-8e59-186e-446a9d949651" [ 895.954183] env[62952]: _type = "HttpNfcLease" [ 895.954183] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 895.962569] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 895.962569] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b14733-2d84-8e59-186e-446a9d949651" [ 895.962569] env[62952]: _type = "HttpNfcLease" [ 895.962569] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 896.047683] env[62952]: INFO nova.compute.manager [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Took 42.09 seconds to build instance. [ 896.117760] env[62952]: DEBUG oslo_vmware.api [None req-8c0eee7c-3ce3-40b3-b6ba-183b3bf25f2c tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263312, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.174363] env[62952]: DEBUG nova.network.neutron [req-00264ebc-52c3-4fbb-9c5e-35d29d56a5bd req-b08b2c41-b22b-4f29-ae3a-3b621fbc2288 service nova] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Updated VIF entry in instance network info cache for port f9a55a9d-b041-4065-b354-8a180d765835. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 896.174804] env[62952]: DEBUG nova.network.neutron [req-00264ebc-52c3-4fbb-9c5e-35d29d56a5bd req-b08b2c41-b22b-4f29-ae3a-3b621fbc2288 service nova] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Updating instance_info_cache with network_info: [{"id": "f9a55a9d-b041-4065-b354-8a180d765835", "address": "fa:16:3e:37:37:40", "network": {"id": "2e11ee50-4303-4f0a-a9fc-464f5f2087b3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "73931ed9ad374fb0870629249c5f6909", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3c995e9-7f2f-420c-880a-d60da6e708ad", "external-id": "nsx-vlan-transportzone-166", "segmentation_id": 166, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9a55a9d-b0", "ovs_interfaceid": "f9a55a9d-b041-4065-b354-8a180d765835", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.214190] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c0800b-8acd-09ba-0617-f8eaaa09092c, 'name': SearchDatastore_Task, 'duration_secs': 0.032942} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.215265] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.215582] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 896.215838] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.216020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.216208] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 896.218063] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21c32148-307b-4b76-a8b6-fe62311d868c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.229558] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 896.229558] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 896.230935] env[62952]: DEBUG nova.network.neutron [req-3ed14ab1-355c-4bca-bad5-799387d10b61 req-60f5b9ec-3bb9-4fe5-98ee-1e08aa4efb4f service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Updated VIF entry in instance network info cache for port 3423511f-8828-4fca-aec2-79e921a53579. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 896.231326] env[62952]: DEBUG nova.network.neutron [req-3ed14ab1-355c-4bca-bad5-799387d10b61 req-60f5b9ec-3bb9-4fe5-98ee-1e08aa4efb4f service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Updating instance_info_cache with network_info: [{"id": "3423511f-8828-4fca-aec2-79e921a53579", "address": "fa:16:3e:3b:bd:29", "network": {"id": "d865ccd7-b8a3-49d7-811f-6d4466bbbf45", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1549996301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4484b1cb78ff4e7baac4b9a5d1e63549", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a5569c57-a27b-4d4d-a519-784be105114f", "external-id": "nsx-vlan-transportzone-640", "segmentation_id": 640, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3423511f-88", "ovs_interfaceid": "3423511f-8828-4fca-aec2-79e921a53579", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.233249] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea4d2432-edfd-463f-8ddf-8faff45ef011 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.242422] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 896.242422] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520edbae-cd7b-a39f-0a28-67c51c8b734c" [ 896.242422] env[62952]: _type = "Task" [ 896.242422] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.251802] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520edbae-cd7b-a39f-0a28-67c51c8b734c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.308785] env[62952]: DEBUG oslo_vmware.api [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': task-1263310, 'name': PowerOnVM_Task, 'duration_secs': 0.544214} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.317363] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 896.317630] env[62952]: INFO nova.compute.manager [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Took 8.82 seconds to spawn the instance on the hypervisor. [ 896.317843] env[62952]: DEBUG nova.compute.manager [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.319387] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04b2c733-4c01-48c0-9f96-4072524f6bde {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.391175] env[62952]: DEBUG oslo_concurrency.lockutils [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquiring lock "refresh_cache-48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.391175] env[62952]: DEBUG oslo_concurrency.lockutils [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquired lock "refresh_cache-48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.391175] env[62952]: DEBUG nova.network.neutron [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 896.413032] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263313, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.462931] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 896.462931] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b14733-2d84-8e59-186e-446a9d949651" [ 896.462931] env[62952]: _type = "HttpNfcLease" [ 896.462931] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 896.463363] env[62952]: DEBUG oslo_vmware.rw_handles [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 896.463363] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b14733-2d84-8e59-186e-446a9d949651" [ 896.463363] env[62952]: _type = "HttpNfcLease" [ 896.463363] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 896.464115] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dca2378-7876-4285-b49d-73606a2f0fc4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.474367] env[62952]: DEBUG oslo_vmware.rw_handles [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529242b2-99e9-5f09-9c4b-de99a85bbbe3/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 896.474567] env[62952]: DEBUG oslo_vmware.rw_handles [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529242b2-99e9-5f09-9c4b-de99a85bbbe3/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 896.549736] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24fc55e7-fc01-4f5b-851f-c95b1f60a106 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "3416eb38-b961-4d48-8b37-8b7b92375dc9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.762s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.584605] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d255e3f7-e16d-48dc-8c08-c1e6ce3ba4b4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.626344] env[62952]: DEBUG oslo_vmware.api [None req-8c0eee7c-3ce3-40b3-b6ba-183b3bf25f2c tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263312, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.646044] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c84fab-bdeb-46ef-b6e5-539931014beb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.655042] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-453b1dfb-b1d2-4f22-904e-ded3892663d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.687178] env[62952]: DEBUG oslo_concurrency.lockutils [req-00264ebc-52c3-4fbb-9c5e-35d29d56a5bd req-b08b2c41-b22b-4f29-ae3a-3b621fbc2288 service nova] Releasing lock "refresh_cache-14f8dbbe-2c4a-4948-bf15-106d5b1b0677" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.687675] env[62952]: DEBUG nova.compute.manager [req-00264ebc-52c3-4fbb-9c5e-35d29d56a5bd req-b08b2c41-b22b-4f29-ae3a-3b621fbc2288 service nova] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Received event network-vif-deleted-95b813f6-3248-4a6a-98b0-033734799b00 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 896.688446] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91b7e36f-1a1f-4eed-bd62-cf5babfa2ab3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.697092] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09de439c-2dbe-4f7f-930a-98ec626fcb9b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.712681] env[62952]: DEBUG nova.compute.provider_tree [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.736986] env[62952]: DEBUG oslo_concurrency.lockutils [req-3ed14ab1-355c-4bca-bad5-799387d10b61 req-60f5b9ec-3bb9-4fe5-98ee-1e08aa4efb4f service nova] Releasing lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.757532] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520edbae-cd7b-a39f-0a28-67c51c8b734c, 'name': SearchDatastore_Task, 'duration_secs': 0.011632} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.760253] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3603e85-0cbc-474f-a76b-51836b3c944a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.769581] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 896.769581] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b00f9f-2f99-bfe2-84b7-f45264a07785" [ 896.769581] env[62952]: _type = "Task" [ 896.769581] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.782304] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b00f9f-2f99-bfe2-84b7-f45264a07785, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.841407] env[62952]: INFO nova.compute.manager [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Took 37.77 seconds to build instance. [ 896.916307] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263313, 'name': ReconfigVM_Task, 'duration_secs': 0.989173} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.916561] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Reconfigured VM instance instance-00000040 to attach disk [datastore2] ba022bea-e357-4541-a573-d1da2eb49b46/ba022bea-e357-4541-a573-d1da2eb49b46.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 896.917348] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-33aa6ae6-bfc0-452a-8ce9-06e8949090c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.923370] env[62952]: DEBUG nova.network.neutron [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 896.927291] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 896.927291] env[62952]: value = "task-1263315" [ 896.927291] env[62952]: _type = "Task" [ 896.927291] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.939412] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263315, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.054797] env[62952]: DEBUG nova.compute.manager [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 897.100256] env[62952]: DEBUG nova.network.neutron [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Updating instance_info_cache with network_info: [{"id": "31694ebf-fc67-4662-98f6-9d99b17eb865", "address": "fa:16:3e:73:3c:7a", "network": {"id": "2e11ee50-4303-4f0a-a9fc-464f5f2087b3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.204", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "73931ed9ad374fb0870629249c5f6909", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3c995e9-7f2f-420c-880a-d60da6e708ad", "external-id": "nsx-vlan-transportzone-166", "segmentation_id": 166, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31694ebf-fc", "ovs_interfaceid": "31694ebf-fc67-4662-98f6-9d99b17eb865", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.125550] env[62952]: DEBUG oslo_vmware.api [None req-8c0eee7c-3ce3-40b3-b6ba-183b3bf25f2c tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263312, 'name': PowerOnVM_Task, 'duration_secs': 1.260218} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.126031] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c0eee7c-3ce3-40b3-b6ba-183b3bf25f2c tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 897.126336] env[62952]: DEBUG nova.compute.manager [None req-8c0eee7c-3ce3-40b3-b6ba-183b3bf25f2c tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.127495] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a6097e-6e94-4c42-be75-5729a8584a0b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.219020] env[62952]: DEBUG nova.scheduler.client.report [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.282230] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b00f9f-2f99-bfe2-84b7-f45264a07785, 'name': SearchDatastore_Task, 'duration_secs': 0.011979} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.282566] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.282844] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 14f8dbbe-2c4a-4948-bf15-106d5b1b0677/14f8dbbe-2c4a-4948-bf15-106d5b1b0677.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 897.283153] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f00e11f7-769c-46b9-b50c-37bbd4054501 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.292150] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 897.292150] env[62952]: value = "task-1263316" [ 897.292150] env[62952]: _type = "Task" [ 897.292150] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.303585] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263316, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.344486] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6950b081-fabd-4c23-a5b4-83ef275fb160 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Lock "49947bde-069b-459c-ae19-ae5f090f535b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.857s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.438858] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263315, 'name': Rename_Task, 'duration_secs': 0.264326} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.439169] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 897.439349] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3387aafa-0891-4fb7-961b-80d9b3fc3f00 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.448799] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 897.448799] env[62952]: value = "task-1263317" [ 897.448799] env[62952]: _type = "Task" [ 897.448799] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.460501] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263317, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.559867] env[62952]: DEBUG nova.compute.manager [req-d6eaa5a5-adc6-483b-83a3-128f44a1792a req-3ebb8903-64ac-4918-88b2-03c5fb2eff8e service nova] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Received event network-vif-plugged-31694ebf-fc67-4662-98f6-9d99b17eb865 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 897.560397] env[62952]: DEBUG oslo_concurrency.lockutils [req-d6eaa5a5-adc6-483b-83a3-128f44a1792a req-3ebb8903-64ac-4918-88b2-03c5fb2eff8e service nova] Acquiring lock "48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.560697] env[62952]: DEBUG oslo_concurrency.lockutils [req-d6eaa5a5-adc6-483b-83a3-128f44a1792a req-3ebb8903-64ac-4918-88b2-03c5fb2eff8e service nova] Lock "48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.560946] env[62952]: DEBUG oslo_concurrency.lockutils [req-d6eaa5a5-adc6-483b-83a3-128f44a1792a req-3ebb8903-64ac-4918-88b2-03c5fb2eff8e service nova] Lock "48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.561317] env[62952]: DEBUG nova.compute.manager [req-d6eaa5a5-adc6-483b-83a3-128f44a1792a req-3ebb8903-64ac-4918-88b2-03c5fb2eff8e service nova] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] No waiting events found dispatching network-vif-plugged-31694ebf-fc67-4662-98f6-9d99b17eb865 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 897.561721] env[62952]: WARNING nova.compute.manager [req-d6eaa5a5-adc6-483b-83a3-128f44a1792a req-3ebb8903-64ac-4918-88b2-03c5fb2eff8e service nova] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Received unexpected event network-vif-plugged-31694ebf-fc67-4662-98f6-9d99b17eb865 for instance with vm_state building and task_state spawning. [ 897.561834] env[62952]: DEBUG nova.compute.manager [req-d6eaa5a5-adc6-483b-83a3-128f44a1792a req-3ebb8903-64ac-4918-88b2-03c5fb2eff8e service nova] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Received event network-changed-31694ebf-fc67-4662-98f6-9d99b17eb865 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 897.562109] env[62952]: DEBUG nova.compute.manager [req-d6eaa5a5-adc6-483b-83a3-128f44a1792a req-3ebb8903-64ac-4918-88b2-03c5fb2eff8e service nova] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Refreshing instance network info cache due to event network-changed-31694ebf-fc67-4662-98f6-9d99b17eb865. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 897.562401] env[62952]: DEBUG oslo_concurrency.lockutils [req-d6eaa5a5-adc6-483b-83a3-128f44a1792a req-3ebb8903-64ac-4918-88b2-03c5fb2eff8e service nova] Acquiring lock "refresh_cache-48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.590175] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.608175] env[62952]: DEBUG oslo_concurrency.lockutils [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Releasing lock "refresh_cache-48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.608658] env[62952]: DEBUG nova.compute.manager [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Instance network_info: |[{"id": "31694ebf-fc67-4662-98f6-9d99b17eb865", "address": "fa:16:3e:73:3c:7a", "network": {"id": "2e11ee50-4303-4f0a-a9fc-464f5f2087b3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.204", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "73931ed9ad374fb0870629249c5f6909", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3c995e9-7f2f-420c-880a-d60da6e708ad", "external-id": "nsx-vlan-transportzone-166", "segmentation_id": 166, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31694ebf-fc", "ovs_interfaceid": "31694ebf-fc67-4662-98f6-9d99b17eb865", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 897.609400] env[62952]: DEBUG oslo_concurrency.lockutils [req-d6eaa5a5-adc6-483b-83a3-128f44a1792a req-3ebb8903-64ac-4918-88b2-03c5fb2eff8e service nova] Acquired lock "refresh_cache-48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.609475] env[62952]: DEBUG nova.network.neutron [req-d6eaa5a5-adc6-483b-83a3-128f44a1792a req-3ebb8903-64ac-4918-88b2-03c5fb2eff8e service nova] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Refreshing network info cache for port 31694ebf-fc67-4662-98f6-9d99b17eb865 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 897.617526] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:3c:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3c995e9-7f2f-420c-880a-d60da6e708ad', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '31694ebf-fc67-4662-98f6-9d99b17eb865', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.640487] env[62952]: DEBUG oslo.service.loopingcall [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.648881] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 897.653308] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-68ff51bb-f63b-47e7-85ac-54f79c5eb090 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.689801] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 897.689801] env[62952]: value = "task-1263318" [ 897.689801] env[62952]: _type = "Task" [ 897.689801] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.705468] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263318, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.723021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.552s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.723648] env[62952]: DEBUG nova.compute.manager [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 897.727230] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.960s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.727497] env[62952]: DEBUG nova.objects.instance [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lazy-loading 'resources' on Instance uuid f906dc47-12cb-46f5-ae5c-53cfcc8765a8 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.809483] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263316, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.847680] env[62952]: DEBUG nova.compute.manager [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 897.963765] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263317, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.980115] env[62952]: DEBUG nova.network.neutron [req-d6eaa5a5-adc6-483b-83a3-128f44a1792a req-3ebb8903-64ac-4918-88b2-03c5fb2eff8e service nova] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Updated VIF entry in instance network info cache for port 31694ebf-fc67-4662-98f6-9d99b17eb865. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 897.980586] env[62952]: DEBUG nova.network.neutron [req-d6eaa5a5-adc6-483b-83a3-128f44a1792a req-3ebb8903-64ac-4918-88b2-03c5fb2eff8e service nova] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Updating instance_info_cache with network_info: [{"id": "31694ebf-fc67-4662-98f6-9d99b17eb865", "address": "fa:16:3e:73:3c:7a", "network": {"id": "2e11ee50-4303-4f0a-a9fc-464f5f2087b3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.204", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "73931ed9ad374fb0870629249c5f6909", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3c995e9-7f2f-420c-880a-d60da6e708ad", "external-id": "nsx-vlan-transportzone-166", "segmentation_id": 166, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31694ebf-fc", "ovs_interfaceid": "31694ebf-fc67-4662-98f6-9d99b17eb865", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.201171] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263318, 'name': CreateVM_Task, 'duration_secs': 0.506597} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.201472] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 898.202216] env[62952]: DEBUG oslo_concurrency.lockutils [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.202456] env[62952]: DEBUG oslo_concurrency.lockutils [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.203008] env[62952]: DEBUG oslo_concurrency.lockutils [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 898.203364] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7190514a-ea39-4bb3-970f-780605339731 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.210133] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 898.210133] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525a5a1b-d628-0317-96c1-776d0dad254e" [ 898.210133] env[62952]: _type = "Task" [ 898.210133] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.218460] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525a5a1b-d628-0317-96c1-776d0dad254e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.221658] env[62952]: DEBUG nova.compute.manager [req-7e6a79e5-a43a-485f-81c7-03b994ab69a5 req-1cd096f1-c2a3-4f69-a137-c644befc4477 service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Received event network-changed-3423511f-8828-4fca-aec2-79e921a53579 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 898.221942] env[62952]: DEBUG nova.compute.manager [req-7e6a79e5-a43a-485f-81c7-03b994ab69a5 req-1cd096f1-c2a3-4f69-a137-c644befc4477 service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Refreshing instance network info cache due to event network-changed-3423511f-8828-4fca-aec2-79e921a53579. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 898.222366] env[62952]: DEBUG oslo_concurrency.lockutils [req-7e6a79e5-a43a-485f-81c7-03b994ab69a5 req-1cd096f1-c2a3-4f69-a137-c644befc4477 service nova] Acquiring lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.222607] env[62952]: DEBUG oslo_concurrency.lockutils [req-7e6a79e5-a43a-485f-81c7-03b994ab69a5 req-1cd096f1-c2a3-4f69-a137-c644befc4477 service nova] Acquired lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.222790] env[62952]: DEBUG nova.network.neutron [req-7e6a79e5-a43a-485f-81c7-03b994ab69a5 req-1cd096f1-c2a3-4f69-a137-c644befc4477 service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Refreshing network info cache for port 3423511f-8828-4fca-aec2-79e921a53579 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 898.234275] env[62952]: DEBUG nova.compute.utils [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 898.237257] env[62952]: DEBUG nova.compute.manager [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Not allocating networking since 'none' was specified. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 898.307656] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263316, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.615399} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.308021] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 14f8dbbe-2c4a-4948-bf15-106d5b1b0677/14f8dbbe-2c4a-4948-bf15-106d5b1b0677.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 898.308474] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 898.308857] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0b7c2f14-2462-44e5-8399-9466e10f51e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.321595] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 898.321595] env[62952]: value = "task-1263319" [ 898.321595] env[62952]: _type = "Task" [ 898.321595] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.337419] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263319, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.375025] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.460728] env[62952]: DEBUG oslo_vmware.api [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263317, 'name': PowerOnVM_Task, 'duration_secs': 0.954465} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.463795] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 898.464112] env[62952]: INFO nova.compute.manager [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Took 9.46 seconds to spawn the instance on the hypervisor. [ 898.464376] env[62952]: DEBUG nova.compute.manager [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 898.465980] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7367115-3de3-4701-aeb5-ceaa0d9071d0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.485298] env[62952]: DEBUG oslo_concurrency.lockutils [req-d6eaa5a5-adc6-483b-83a3-128f44a1792a req-3ebb8903-64ac-4918-88b2-03c5fb2eff8e service nova] Releasing lock "refresh_cache-48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.635102] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f15f28-94f1-4b40-9632-13297624149c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.641809] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c70425-7443-4c06-8beb-12d91e769c62 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.684867] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f13f5899-3072-4b4c-a03e-36b97f0f8edd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.694664] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a51737-46b0-4bf6-84e8-7053a25a0fc3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.710619] env[62952]: DEBUG nova.compute.provider_tree [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.723359] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525a5a1b-d628-0317-96c1-776d0dad254e, 'name': SearchDatastore_Task, 'duration_secs': 0.01371} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.724487] env[62952]: DEBUG oslo_concurrency.lockutils [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.724735] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.725355] env[62952]: DEBUG oslo_concurrency.lockutils [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.725355] env[62952]: DEBUG oslo_concurrency.lockutils [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.725355] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.725826] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fd6f3e9c-2066-4dfd-9554-f6edf80c28fd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.739829] env[62952]: DEBUG nova.compute.manager [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 898.742919] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.743130] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 898.744604] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ef01249-1980-4f2e-ab07-b03dd3c2d9e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.751350] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 898.751350] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52179f91-0070-a62f-5623-07e5ce45699c" [ 898.751350] env[62952]: _type = "Task" [ 898.751350] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.762781] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52179f91-0070-a62f-5623-07e5ce45699c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.832444] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263319, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.112101} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.832794] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 898.839050] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-814b9b90-2b04-4c3e-a2b3-44382bb3c461 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.871421] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 14f8dbbe-2c4a-4948-bf15-106d5b1b0677/14f8dbbe-2c4a-4948-bf15-106d5b1b0677.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 898.874590] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5534702-ce20-48d5-b201-11adb8ac0919 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.899918] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 898.899918] env[62952]: value = "task-1263320" [ 898.899918] env[62952]: _type = "Task" [ 898.899918] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.909969] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263320, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.988029] env[62952]: INFO nova.compute.manager [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Took 34.13 seconds to build instance. [ 899.000668] env[62952]: DEBUG nova.network.neutron [req-7e6a79e5-a43a-485f-81c7-03b994ab69a5 req-1cd096f1-c2a3-4f69-a137-c644befc4477 service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Updated VIF entry in instance network info cache for port 3423511f-8828-4fca-aec2-79e921a53579. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 899.003130] env[62952]: DEBUG nova.network.neutron [req-7e6a79e5-a43a-485f-81c7-03b994ab69a5 req-1cd096f1-c2a3-4f69-a137-c644befc4477 service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Updating instance_info_cache with network_info: [{"id": "3423511f-8828-4fca-aec2-79e921a53579", "address": "fa:16:3e:3b:bd:29", "network": {"id": "d865ccd7-b8a3-49d7-811f-6d4466bbbf45", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1549996301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4484b1cb78ff4e7baac4b9a5d1e63549", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a5569c57-a27b-4d4d-a519-784be105114f", "external-id": "nsx-vlan-transportzone-640", "segmentation_id": 640, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3423511f-88", "ovs_interfaceid": "3423511f-8828-4fca-aec2-79e921a53579", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.153360] env[62952]: INFO nova.compute.manager [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Rebuilding instance [ 899.201968] env[62952]: DEBUG nova.compute.manager [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 899.203032] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9f9bbe-0af9-423e-94f2-be355c8b1be2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.216886] env[62952]: DEBUG nova.scheduler.client.report [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 899.262623] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52179f91-0070-a62f-5623-07e5ce45699c, 'name': SearchDatastore_Task, 'duration_secs': 0.014382} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.263608] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-336fd3aa-8c17-4ecd-b836-4d28564c2b13 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.269940] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 899.269940] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521bc4d4-19ba-9174-5812-18160bb8713e" [ 899.269940] env[62952]: _type = "Task" [ 899.269940] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.279568] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521bc4d4-19ba-9174-5812-18160bb8713e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.411354] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263320, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.489637] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53fefa8e-552c-4b9c-951c-bcf92b595c9c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "ba022bea-e357-4541-a573-d1da2eb49b46" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.039s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.506477] env[62952]: DEBUG oslo_concurrency.lockutils [req-7e6a79e5-a43a-485f-81c7-03b994ab69a5 req-1cd096f1-c2a3-4f69-a137-c644befc4477 service nova] Releasing lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.526476] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Acquiring lock "49947bde-069b-459c-ae19-ae5f090f535b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.526753] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Lock "49947bde-069b-459c-ae19-ae5f090f535b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.526969] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Acquiring lock "49947bde-069b-459c-ae19-ae5f090f535b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.527175] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Lock "49947bde-069b-459c-ae19-ae5f090f535b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.527346] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Lock "49947bde-069b-459c-ae19-ae5f090f535b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.529621] env[62952]: INFO nova.compute.manager [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Terminating instance [ 899.531657] env[62952]: DEBUG nova.compute.manager [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 899.531849] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 899.532750] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8baa90e0-1691-4f96-915c-0fdbd499ba67 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.541104] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 899.541446] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c409cec-75c8-4e99-a18b-d8a2f82a6c66 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.550739] env[62952]: DEBUG oslo_vmware.api [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Waiting for the task: (returnval){ [ 899.550739] env[62952]: value = "task-1263321" [ 899.550739] env[62952]: _type = "Task" [ 899.550739] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.560514] env[62952]: DEBUG oslo_vmware.api [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': task-1263321, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.723063] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 899.724073] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.997s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.726854] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cf6e8cab-b118-4714-a40e-ff17c300cc9b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.728706] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.089s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.730501] env[62952]: INFO nova.compute.claims [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 899.741691] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 899.741691] env[62952]: value = "task-1263322" [ 899.741691] env[62952]: _type = "Task" [ 899.741691] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.761381] env[62952]: INFO nova.scheduler.client.report [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Deleted allocations for instance f906dc47-12cb-46f5-ae5c-53cfcc8765a8 [ 899.763642] env[62952]: DEBUG nova.compute.manager [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 899.782302] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263322, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.790748] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521bc4d4-19ba-9174-5812-18160bb8713e, 'name': SearchDatastore_Task, 'duration_secs': 0.013893} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.792250] env[62952]: DEBUG oslo_concurrency.lockutils [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.792915] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5/48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 899.793855] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c03b4cd2-cb11-4ed8-a446-298980936be1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.802596] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 899.802596] env[62952]: value = "task-1263323" [ 899.802596] env[62952]: _type = "Task" [ 899.802596] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.807566] env[62952]: DEBUG nova.virt.hardware [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 899.807960] env[62952]: DEBUG nova.virt.hardware [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 899.808206] env[62952]: DEBUG nova.virt.hardware [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 899.808465] env[62952]: DEBUG nova.virt.hardware [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 899.809032] env[62952]: DEBUG nova.virt.hardware [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 899.809032] env[62952]: DEBUG nova.virt.hardware [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 899.809180] env[62952]: DEBUG nova.virt.hardware [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 899.809357] env[62952]: DEBUG nova.virt.hardware [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 899.809590] env[62952]: DEBUG nova.virt.hardware [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 899.809813] env[62952]: DEBUG nova.virt.hardware [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 899.810177] env[62952]: DEBUG nova.virt.hardware [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 899.816042] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7615fd66-c555-4c5d-9b42-8563d3342bde {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.829670] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263323, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.834508] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed0d84c-9847-4eb5-83a6-7352551e04a3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.853470] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 899.859146] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Creating folder: Project (fe25062529174c5490daa41cc5f3be10). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 899.859929] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a331576d-23a3-4560-80b1-3a735f01463e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.874341] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Created folder: Project (fe25062529174c5490daa41cc5f3be10) in parent group-v271811. [ 899.874608] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Creating folder: Instances. Parent ref: group-v271889. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 899.874905] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5e23a808-dd4e-4f68-b589-07759cbafbab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.887635] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Created folder: Instances in parent group-v271889. [ 899.888019] env[62952]: DEBUG oslo.service.loopingcall [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.888299] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 899.888618] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a81360db-6c07-44e5-baaa-d3923e5fc4bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.914055] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263320, 'name': ReconfigVM_Task, 'duration_secs': 0.514702} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.915510] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 14f8dbbe-2c4a-4948-bf15-106d5b1b0677/14f8dbbe-2c4a-4948-bf15-106d5b1b0677.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 899.916244] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 899.916244] env[62952]: value = "task-1263326" [ 899.916244] env[62952]: _type = "Task" [ 899.916244] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.916714] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c74b512-5e00-496b-84ea-f418f2b2b6c3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.929492] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263326, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.931017] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 899.931017] env[62952]: value = "task-1263327" [ 899.931017] env[62952]: _type = "Task" [ 899.931017] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.941247] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263327, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.065302] env[62952]: DEBUG oslo_vmware.api [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': task-1263321, 'name': PowerOffVM_Task, 'duration_secs': 0.282961} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.065758] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 900.066038] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 900.066371] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9a5e5b29-46a5-42df-9bd0-5a4796fb3f41 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.105190] env[62952]: DEBUG oslo_concurrency.lockutils [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Acquiring lock "67b729be-fb66-4ed1-bbea-e62216d460d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.105190] env[62952]: DEBUG oslo_concurrency.lockutils [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Lock "67b729be-fb66-4ed1-bbea-e62216d460d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.105584] env[62952]: DEBUG oslo_concurrency.lockutils [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Acquiring lock "67b729be-fb66-4ed1-bbea-e62216d460d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.105874] env[62952]: DEBUG oslo_concurrency.lockutils [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Lock "67b729be-fb66-4ed1-bbea-e62216d460d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.106176] env[62952]: DEBUG oslo_concurrency.lockutils [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Lock "67b729be-fb66-4ed1-bbea-e62216d460d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.108909] env[62952]: INFO nova.compute.manager [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Terminating instance [ 900.111698] env[62952]: DEBUG nova.compute.manager [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 900.111906] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 900.112799] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac3b4338-567c-41bf-93c6-164ee57c1110 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.125189] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 900.125548] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5074bca1-40d3-487d-81fd-c5f9b179149b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.136495] env[62952]: DEBUG oslo_vmware.api [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 900.136495] env[62952]: value = "task-1263329" [ 900.136495] env[62952]: _type = "Task" [ 900.136495] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.148361] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 900.148566] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 900.148813] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Deleting the datastore file [datastore2] 49947bde-069b-459c-ae19-ae5f090f535b {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.149781] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63c32504-5b8d-47f8-a937-931ca30c2471 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.154898] env[62952]: DEBUG oslo_vmware.api [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263329, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.161624] env[62952]: DEBUG oslo_vmware.api [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Waiting for the task: (returnval){ [ 900.161624] env[62952]: value = "task-1263330" [ 900.161624] env[62952]: _type = "Task" [ 900.161624] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.172270] env[62952]: DEBUG oslo_vmware.api [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': task-1263330, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.254992] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263322, 'name': PowerOffVM_Task, 'duration_secs': 0.255421} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.256311] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 900.256562] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 900.257946] env[62952]: DEBUG nova.compute.manager [req-c5f08687-4319-46a3-bc5d-bfb45f73a638 req-e54f8876-59c6-40dd-ba63-391d10c7a06c service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Received event network-changed-3423511f-8828-4fca-aec2-79e921a53579 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 900.258165] env[62952]: DEBUG nova.compute.manager [req-c5f08687-4319-46a3-bc5d-bfb45f73a638 req-e54f8876-59c6-40dd-ba63-391d10c7a06c service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Refreshing instance network info cache due to event network-changed-3423511f-8828-4fca-aec2-79e921a53579. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 900.258393] env[62952]: DEBUG oslo_concurrency.lockutils [req-c5f08687-4319-46a3-bc5d-bfb45f73a638 req-e54f8876-59c6-40dd-ba63-391d10c7a06c service nova] Acquiring lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.258543] env[62952]: DEBUG oslo_concurrency.lockutils [req-c5f08687-4319-46a3-bc5d-bfb45f73a638 req-e54f8876-59c6-40dd-ba63-391d10c7a06c service nova] Acquired lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.258709] env[62952]: DEBUG nova.network.neutron [req-c5f08687-4319-46a3-bc5d-bfb45f73a638 req-e54f8876-59c6-40dd-ba63-391d10c7a06c service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Refreshing network info cache for port 3423511f-8828-4fca-aec2-79e921a53579 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 900.260483] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd42a4b-45da-4d7f-bd51-f2da3160885d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.272251] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 900.273091] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4d994be2-cfad-4628-80f1-618dcf081d99 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.277523] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42951fce-0aa8-4e14-91d8-3ba9d6a9c9f6 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "f906dc47-12cb-46f5-ae5c-53cfcc8765a8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.211s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.315870] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263323, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.373151] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 900.373570] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 900.373812] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleting the datastore file [datastore1] 3416eb38-b961-4d48-8b37-8b7b92375dc9 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.374228] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7896163d-dffd-4779-8f07-890ce1a78ae9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.384084] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 900.384084] env[62952]: value = "task-1263332" [ 900.384084] env[62952]: _type = "Task" [ 900.384084] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.397264] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263332, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.428998] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263326, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.449054] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263327, 'name': Rename_Task, 'duration_secs': 0.188218} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.451035] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 900.451035] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-123b5513-2f46-4c70-80a6-67bcd152f0ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.459527] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 900.459527] env[62952]: value = "task-1263333" [ 900.459527] env[62952]: _type = "Task" [ 900.459527] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.470762] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263333, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.540113] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "ba022bea-e357-4541-a573-d1da2eb49b46" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.540498] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "ba022bea-e357-4541-a573-d1da2eb49b46" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.540777] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "ba022bea-e357-4541-a573-d1da2eb49b46-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.541050] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "ba022bea-e357-4541-a573-d1da2eb49b46-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.541280] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "ba022bea-e357-4541-a573-d1da2eb49b46-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.544875] env[62952]: INFO nova.compute.manager [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Terminating instance [ 900.548075] env[62952]: DEBUG nova.compute.manager [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 900.548075] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 900.548653] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-319cdb08-ea1d-4f19-8554-6345db32f2cb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.558091] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 900.558407] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a742f94e-75ad-4a34-9d74-7947245ba81f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.566601] env[62952]: DEBUG oslo_vmware.api [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 900.566601] env[62952]: value = "task-1263334" [ 900.566601] env[62952]: _type = "Task" [ 900.566601] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.576232] env[62952]: DEBUG oslo_vmware.api [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263334, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.648393] env[62952]: DEBUG oslo_vmware.api [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263329, 'name': PowerOffVM_Task, 'duration_secs': 0.291747} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.648824] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 900.648974] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 900.649293] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8982ea0d-e504-49fc-8bad-f49582071c87 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.674271] env[62952]: DEBUG oslo_vmware.api [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Task: {'id': task-1263330, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.402069} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.674554] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 900.674744] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 900.674925] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 900.675164] env[62952]: INFO nova.compute.manager [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Took 1.14 seconds to destroy the instance on the hypervisor. [ 900.675440] env[62952]: DEBUG oslo.service.loopingcall [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.675644] env[62952]: DEBUG nova.compute.manager [-] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 900.675741] env[62952]: DEBUG nova.network.neutron [-] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 900.771042] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 900.771239] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 900.771566] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Deleting the datastore file [datastore1] 67b729be-fb66-4ed1-bbea-e62216d460d5 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.772530] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b2297b1d-d1b8-4ae8-937c-9fc8e70ae3cd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.781561] env[62952]: DEBUG oslo_vmware.api [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for the task: (returnval){ [ 900.781561] env[62952]: value = "task-1263336" [ 900.781561] env[62952]: _type = "Task" [ 900.781561] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.796757] env[62952]: DEBUG oslo_vmware.api [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263336, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.820552] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263323, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.590607} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.820991] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5/48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 900.821241] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 900.821574] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-300543bb-f03d-48e3-bfba-aaab227e3217 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.830155] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 900.830155] env[62952]: value = "task-1263337" [ 900.830155] env[62952]: _type = "Task" [ 900.830155] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.842815] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263337, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.903983] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263332, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.264734} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.904347] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 900.904966] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 900.904966] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 900.938992] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "58e703ea-7ccf-4114-a5a2-49533390ed37" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.939309] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "58e703ea-7ccf-4114-a5a2-49533390ed37" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.939542] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "58e703ea-7ccf-4114-a5a2-49533390ed37-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.943027] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "58e703ea-7ccf-4114-a5a2-49533390ed37-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.943027] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "58e703ea-7ccf-4114-a5a2-49533390ed37-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.943027] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263326, 'name': CreateVM_Task, 'duration_secs': 0.651236} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.943027] env[62952]: INFO nova.compute.manager [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Terminating instance [ 900.943715] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 900.944228] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.944341] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.944698] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 900.944966] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f319f1c-58e5-4770-9956-44dea9e42d0c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.951986] env[62952]: DEBUG nova.compute.manager [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 900.951986] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 900.951986] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eacc1f34-8e45-42de-aa56-90a8542467ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.960316] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 900.961789] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a636af9b-c3ac-4625-a99d-eef69652f921 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.963639] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 900.963639] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523214b2-d321-2e8b-fabb-f7e68a6defab" [ 900.963639] env[62952]: _type = "Task" [ 900.963639] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.984259] env[62952]: DEBUG oslo_vmware.api [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 900.984259] env[62952]: value = "task-1263338" [ 900.984259] env[62952]: _type = "Task" [ 900.984259] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.987999] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263333, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.996892] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523214b2-d321-2e8b-fabb-f7e68a6defab, 'name': SearchDatastore_Task, 'duration_secs': 0.018431} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.997175] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.997474] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 900.997723] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.998840] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.998840] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 901.002549] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-302c7bd9-3123-466c-8265-6fd3cfb02447 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.010611] env[62952]: DEBUG oslo_vmware.api [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263338, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.022852] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 901.022936] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 901.024118] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9224af0-9e9f-40cd-919d-88b648ab0bd3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.034259] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 901.034259] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]528d7f62-c6c3-1c99-76fb-cdd47ff62f10" [ 901.034259] env[62952]: _type = "Task" [ 901.034259] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.044250] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]528d7f62-c6c3-1c99-76fb-cdd47ff62f10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.077320] env[62952]: DEBUG oslo_vmware.api [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263334, 'name': PowerOffVM_Task, 'duration_secs': 0.309476} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.082370] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 901.082575] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 901.083073] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dbc8fedd-5c83-4b16-b56d-c81800161e68 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.129872] env[62952]: DEBUG nova.network.neutron [req-c5f08687-4319-46a3-bc5d-bfb45f73a638 req-e54f8876-59c6-40dd-ba63-391d10c7a06c service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Updated VIF entry in instance network info cache for port 3423511f-8828-4fca-aec2-79e921a53579. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 901.130281] env[62952]: DEBUG nova.network.neutron [req-c5f08687-4319-46a3-bc5d-bfb45f73a638 req-e54f8876-59c6-40dd-ba63-391d10c7a06c service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Updating instance_info_cache with network_info: [{"id": "3423511f-8828-4fca-aec2-79e921a53579", "address": "fa:16:3e:3b:bd:29", "network": {"id": "d865ccd7-b8a3-49d7-811f-6d4466bbbf45", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1549996301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4484b1cb78ff4e7baac4b9a5d1e63549", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a5569c57-a27b-4d4d-a519-784be105114f", "external-id": "nsx-vlan-transportzone-640", "segmentation_id": 640, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3423511f-88", "ovs_interfaceid": "3423511f-8828-4fca-aec2-79e921a53579", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.134702] env[62952]: DEBUG nova.compute.manager [req-ef4f3b80-67eb-415b-8d5e-705aed8f9b61 req-023c968f-9500-4795-9ae3-1cb86e9a68a8 service nova] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Received event network-vif-deleted-968b1d86-0e93-4f9a-9003-392cfa822d2c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 901.134955] env[62952]: INFO nova.compute.manager [req-ef4f3b80-67eb-415b-8d5e-705aed8f9b61 req-023c968f-9500-4795-9ae3-1cb86e9a68a8 service nova] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Neutron deleted interface 968b1d86-0e93-4f9a-9003-392cfa822d2c; detaching it from the instance and deleting it from the info cache [ 901.135091] env[62952]: DEBUG nova.network.neutron [req-ef4f3b80-67eb-415b-8d5e-705aed8f9b61 req-023c968f-9500-4795-9ae3-1cb86e9a68a8 service nova] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.172766] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 901.173009] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 901.173204] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleting the datastore file [datastore2] ba022bea-e357-4541-a573-d1da2eb49b46 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 901.176009] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-47c1d89b-238f-44f1-ad02-4e63d0c294e8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.184704] env[62952]: DEBUG oslo_vmware.api [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 901.184704] env[62952]: value = "task-1263340" [ 901.184704] env[62952]: _type = "Task" [ 901.184704] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.190310] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b01502-3630-48ba-90ba-7eb56a7b88cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.200019] env[62952]: DEBUG oslo_vmware.api [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263340, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.205027] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc755bed-e40e-47c4-abaa-bab712bddbd6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.238188] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad81a554-7d4c-4ee6-9891-e3a2115829ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.247861] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5a30de8-5b18-4f67-8c81-6536b6b9cc33 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.271447] env[62952]: DEBUG nova.compute.provider_tree [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.293287] env[62952]: DEBUG oslo_vmware.api [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Task: {'id': task-1263336, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.375677} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.294050] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.294675] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 901.294675] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 901.294826] env[62952]: INFO nova.compute.manager [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Took 1.18 seconds to destroy the instance on the hypervisor. [ 901.295078] env[62952]: DEBUG oslo.service.loopingcall [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.295448] env[62952]: DEBUG nova.compute.manager [-] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 901.295608] env[62952]: DEBUG nova.network.neutron [-] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 901.340679] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263337, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073457} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.340934] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 901.341875] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72898e7-0151-45fa-a716-5074a2c02db4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.365830] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5/48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 901.368427] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-027227b1-1000-4fb5-808a-7239bc32580d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.393380] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 901.393380] env[62952]: value = "task-1263341" [ 901.393380] env[62952]: _type = "Task" [ 901.393380] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.402705] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263341, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.472621] env[62952]: DEBUG oslo_vmware.api [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263333, 'name': PowerOnVM_Task, 'duration_secs': 0.53232} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.473038] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 901.473312] env[62952]: INFO nova.compute.manager [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Took 8.91 seconds to spawn the instance on the hypervisor. [ 901.473604] env[62952]: DEBUG nova.compute.manager [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 901.474816] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6abdf9d2-2ffe-4e28-8cf3-43cce99867ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.485287] env[62952]: DEBUG nova.network.neutron [-] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.498523] env[62952]: DEBUG oslo_vmware.api [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263338, 'name': PowerOffVM_Task, 'duration_secs': 0.247566} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.499259] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 901.499259] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 901.499380] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ef9316c4-21d1-4479-8ee0-31a7808bd891 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.545968] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]528d7f62-c6c3-1c99-76fb-cdd47ff62f10, 'name': SearchDatastore_Task, 'duration_secs': 0.01312} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.547115] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30daebfc-63b8-4015-93b7-8e9ff2bf4136 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.554309] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 901.554309] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b32c48-8ee0-9b6f-e586-9acb5d80af58" [ 901.554309] env[62952]: _type = "Task" [ 901.554309] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.564672] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b32c48-8ee0-9b6f-e586-9acb5d80af58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.598209] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 901.598209] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 901.598209] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Deleting the datastore file [datastore2] 58e703ea-7ccf-4114-a5a2-49533390ed37 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 901.598209] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b58ae01c-25fc-4b25-b61c-ec76bbc389f8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.606186] env[62952]: DEBUG oslo_vmware.api [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 901.606186] env[62952]: value = "task-1263343" [ 901.606186] env[62952]: _type = "Task" [ 901.606186] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.615753] env[62952]: DEBUG oslo_vmware.api [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263343, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.634083] env[62952]: DEBUG oslo_concurrency.lockutils [req-c5f08687-4319-46a3-bc5d-bfb45f73a638 req-e54f8876-59c6-40dd-ba63-391d10c7a06c service nova] Releasing lock "refresh_cache-67b729be-fb66-4ed1-bbea-e62216d460d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.642033] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ad1e9307-0ddd-4ad7-98a5-5107378c1ebb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.656022] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c39c4369-d2c6-4cd6-8823-7b4e44f91fea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.691399] env[62952]: DEBUG nova.compute.manager [req-ef4f3b80-67eb-415b-8d5e-705aed8f9b61 req-023c968f-9500-4795-9ae3-1cb86e9a68a8 service nova] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Detach interface failed, port_id=968b1d86-0e93-4f9a-9003-392cfa822d2c, reason: Instance 49947bde-069b-459c-ae19-ae5f090f535b could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 901.701098] env[62952]: DEBUG oslo_vmware.api [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263340, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.295671} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.701378] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.701553] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 901.701739] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 901.701914] env[62952]: INFO nova.compute.manager [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Took 1.15 seconds to destroy the instance on the hypervisor. [ 901.702204] env[62952]: DEBUG oslo.service.loopingcall [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.702378] env[62952]: DEBUG nova.compute.manager [-] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 901.702477] env[62952]: DEBUG nova.network.neutron [-] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 901.776217] env[62952]: DEBUG nova.scheduler.client.report [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.906154] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263341, 'name': ReconfigVM_Task, 'duration_secs': 0.510913} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.906627] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5/48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 901.907441] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5c6a7c8a-a239-4de1-b0d9-2339c63aa583 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.919761] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 901.919761] env[62952]: value = "task-1263344" [ 901.919761] env[62952]: _type = "Task" [ 901.919761] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.930220] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263344, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.948527] env[62952]: DEBUG nova.virt.hardware [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.948805] env[62952]: DEBUG nova.virt.hardware [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.949320] env[62952]: DEBUG nova.virt.hardware [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.949320] env[62952]: DEBUG nova.virt.hardware [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.949320] env[62952]: DEBUG nova.virt.hardware [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.949557] env[62952]: DEBUG nova.virt.hardware [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.949679] env[62952]: DEBUG nova.virt.hardware [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.949838] env[62952]: DEBUG nova.virt.hardware [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.950018] env[62952]: DEBUG nova.virt.hardware [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.950238] env[62952]: DEBUG nova.virt.hardware [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.950416] env[62952]: DEBUG nova.virt.hardware [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.951753] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4081f8de-c2a2-4bd9-8379-0cc1a699c25a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.960613] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6888593d-23e8-46e8-b284-e6b48acda923 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.977238] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:97:d6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5ad8894f-e240-4013-8272-4e79daea0751', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd4ed51c5-2a91-40ed-8ce9-62f6109ab3b8', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.985499] env[62952]: DEBUG oslo.service.loopingcall [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.985691] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 901.986071] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6ace7c6b-8d37-40ae-b91c-24894bfd2776 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.009895] env[62952]: INFO nova.compute.manager [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Took 34.90 seconds to build instance. [ 902.009895] env[62952]: INFO nova.compute.manager [-] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Took 1.33 seconds to deallocate network for instance. [ 902.018389] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 902.018389] env[62952]: value = "task-1263345" [ 902.018389] env[62952]: _type = "Task" [ 902.018389] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.027663] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263345, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.067403] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b32c48-8ee0-9b6f-e586-9acb5d80af58, 'name': SearchDatastore_Task, 'duration_secs': 0.017276} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.067738] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.068077] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 86ec7b3e-34d6-45e8-87b2-c8550f035ed0/86ec7b3e-34d6-45e8-87b2-c8550f035ed0.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 902.068383] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ea913390-e881-467e-9c8f-d81721a61c84 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.082284] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 902.082284] env[62952]: value = "task-1263346" [ 902.082284] env[62952]: _type = "Task" [ 902.082284] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.091892] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263346, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.117531] env[62952]: DEBUG oslo_vmware.api [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263343, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.277289} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.117856] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 902.118093] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 902.118320] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 902.118555] env[62952]: INFO nova.compute.manager [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Took 1.17 seconds to destroy the instance on the hypervisor. [ 902.118846] env[62952]: DEBUG oslo.service.loopingcall [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 902.119092] env[62952]: DEBUG nova.compute.manager [-] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 902.119195] env[62952]: DEBUG nova.network.neutron [-] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 902.170043] env[62952]: DEBUG nova.network.neutron [-] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.284305] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.555s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.285074] env[62952]: DEBUG nova.compute.manager [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 902.288212] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.585s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.289485] env[62952]: INFO nova.compute.claims [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 902.298289] env[62952]: DEBUG nova.compute.manager [req-b4811fca-4e6e-433a-8cb5-85f7fee1bcf9 req-0d203700-0e63-4016-8ae4-b67b1a7b906e service nova] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Received event network-vif-deleted-3423511f-8828-4fca-aec2-79e921a53579 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 902.298490] env[62952]: DEBUG nova.compute.manager [req-b4811fca-4e6e-433a-8cb5-85f7fee1bcf9 req-0d203700-0e63-4016-8ae4-b67b1a7b906e service nova] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Received event network-vif-deleted-ac76c133-49eb-4a8a-afe8-015a43840974 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 902.298653] env[62952]: INFO nova.compute.manager [req-b4811fca-4e6e-433a-8cb5-85f7fee1bcf9 req-0d203700-0e63-4016-8ae4-b67b1a7b906e service nova] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Neutron deleted interface ac76c133-49eb-4a8a-afe8-015a43840974; detaching it from the instance and deleting it from the info cache [ 902.298831] env[62952]: DEBUG nova.network.neutron [req-b4811fca-4e6e-433a-8cb5-85f7fee1bcf9 req-0d203700-0e63-4016-8ae4-b67b1a7b906e service nova] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.433171] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263344, 'name': Rename_Task, 'duration_secs': 0.285114} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.435442] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 902.436087] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f2b48d9-7bb1-4372-9a93-ff05d7ece6ee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.444784] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 902.444784] env[62952]: value = "task-1263347" [ 902.444784] env[62952]: _type = "Task" [ 902.444784] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.455339] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263347, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.517697] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.517966] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6ca5a05-9ed1-4b8b-b6cd-5acc5a83fc27 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "14f8dbbe-2c4a-4948-bf15-106d5b1b0677" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.383s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.531787] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263345, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.586447] env[62952]: DEBUG nova.network.neutron [-] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.592893] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263346, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.671766] env[62952]: INFO nova.compute.manager [-] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Took 1.38 seconds to deallocate network for instance. [ 902.795168] env[62952]: DEBUG nova.compute.utils [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 902.801253] env[62952]: DEBUG nova.compute.manager [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Not allocating networking since 'none' was specified. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 902.801954] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d0918223-8057-4c44-a67d-91adc65030f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.816272] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f64c8e-9be2-46f2-9655-dd1e2efa89c9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.852996] env[62952]: DEBUG nova.compute.manager [req-b4811fca-4e6e-433a-8cb5-85f7fee1bcf9 req-0d203700-0e63-4016-8ae4-b67b1a7b906e service nova] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Detach interface failed, port_id=ac76c133-49eb-4a8a-afe8-015a43840974, reason: Instance ba022bea-e357-4541-a573-d1da2eb49b46 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 902.938041] env[62952]: DEBUG nova.network.neutron [-] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.957150] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263347, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.030892] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263345, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.092122] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263346, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.723302} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.092528] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 86ec7b3e-34d6-45e8-87b2-c8550f035ed0/86ec7b3e-34d6-45e8-87b2-c8550f035ed0.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 903.092678] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.092971] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d42f34eb-a44a-474f-968f-1aad3d57ef84 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.097287] env[62952]: INFO nova.compute.manager [-] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Took 1.39 seconds to deallocate network for instance. [ 903.107518] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 903.107518] env[62952]: value = "task-1263348" [ 903.107518] env[62952]: _type = "Task" [ 903.107518] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.118480] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263348, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.179621] env[62952]: DEBUG oslo_concurrency.lockutils [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.298307] env[62952]: DEBUG nova.compute.manager [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 903.439941] env[62952]: INFO nova.compute.manager [-] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Took 1.32 seconds to deallocate network for instance. [ 903.460597] env[62952]: DEBUG oslo_vmware.api [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263347, 'name': PowerOnVM_Task, 'duration_secs': 0.834426} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.460844] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 903.461079] env[62952]: INFO nova.compute.manager [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Took 8.30 seconds to spawn the instance on the hypervisor. [ 903.461272] env[62952]: DEBUG nova.compute.manager [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.462182] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60dab5d5-66b5-4d14-8ab2-7253a9d8b3c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.535909] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263345, 'name': CreateVM_Task, 'duration_secs': 1.284508} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.536113] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 903.536967] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.537216] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.537532] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 903.537838] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abcd9329-07d4-498d-8222-fc701c419eae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.543772] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 903.543772] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e8417d-8eeb-ec85-7f0a-ef4ce569575b" [ 903.543772] env[62952]: _type = "Task" [ 903.543772] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.557061] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e8417d-8eeb-ec85-7f0a-ef4ce569575b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.605051] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.621779] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263348, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.128322} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.622214] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 903.622890] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f991ec-be63-4c3c-a23a-80dbf19dc187 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.646125] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 86ec7b3e-34d6-45e8-87b2-c8550f035ed0/86ec7b3e-34d6-45e8-87b2-c8550f035ed0.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 903.649479] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c4f369cb-8f7f-43aa-88cf-e979f4005862 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.675856] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 903.675856] env[62952]: value = "task-1263349" [ 903.675856] env[62952]: _type = "Task" [ 903.675856] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.681846] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f073806d-77ff-48dc-95e4-dfd144e26524 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.691514] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263349, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.694934] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e8591c5-a25d-453c-8290-a5135fbc085f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.732292] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a02efbe9-0428-41b3-b82f-1276eb8669be {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.741991] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e36dafd0-6b1e-4613-bd49-b6479a9ca078 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.762890] env[62952]: DEBUG nova.compute.provider_tree [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.946696] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.985707] env[62952]: INFO nova.compute.manager [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Took 35.72 seconds to build instance. [ 904.057543] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e8417d-8eeb-ec85-7f0a-ef4ce569575b, 'name': SearchDatastore_Task, 'duration_secs': 0.016178} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.057923] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.058227] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 904.058509] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.058706] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.058947] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 904.059279] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ccbf6043-b302-459a-a5e4-333f0fda6645 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.070914] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 904.071205] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 904.072289] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1105af24-79ab-46ca-ac47-ed1ae95948bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.080381] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 904.080381] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526dca42-d72e-09d4-cc41-d3eed3c3917d" [ 904.080381] env[62952]: _type = "Task" [ 904.080381] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.092039] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526dca42-d72e-09d4-cc41-d3eed3c3917d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.187700] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263349, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.266553] env[62952]: DEBUG nova.scheduler.client.report [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 904.309707] env[62952]: DEBUG nova.compute.manager [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 904.336193] env[62952]: DEBUG nova.virt.hardware [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 904.336471] env[62952]: DEBUG nova.virt.hardware [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 904.336629] env[62952]: DEBUG nova.virt.hardware [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 904.336808] env[62952]: DEBUG nova.virt.hardware [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 904.336956] env[62952]: DEBUG nova.virt.hardware [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 904.337153] env[62952]: DEBUG nova.virt.hardware [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 904.337327] env[62952]: DEBUG nova.virt.hardware [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 904.337492] env[62952]: DEBUG nova.virt.hardware [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 904.337663] env[62952]: DEBUG nova.virt.hardware [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 904.337828] env[62952]: DEBUG nova.virt.hardware [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 904.338091] env[62952]: DEBUG nova.virt.hardware [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 904.338894] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27d7b1a-d6a3-46e7-9654-36414c4cf0de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.347929] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4135a7b5-2c4a-48f4-a70c-f896683a3827 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.363652] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 904.369303] env[62952]: DEBUG oslo.service.loopingcall [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.369964] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 904.370225] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-971d3447-6968-4756-b6a0-4b2aab7c6cb3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.386301] env[62952]: DEBUG nova.compute.manager [req-6ad9cb87-698c-44c0-ad50-d53cc76458e5 req-4a24a8a2-0b9e-4dbc-b14a-87e37ce9e70f service nova] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Received event network-vif-deleted-ef78c1ce-2d7f-4e00-87a2-95a93cfc3f37 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 904.393956] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 904.393956] env[62952]: value = "task-1263350" [ 904.393956] env[62952]: _type = "Task" [ 904.393956] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.403339] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263350, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.487944] env[62952]: DEBUG oslo_concurrency.lockutils [None req-194d24de-1e2c-49b2-846c-c6b80ae63e8b tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.192s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.594291] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526dca42-d72e-09d4-cc41-d3eed3c3917d, 'name': SearchDatastore_Task, 'duration_secs': 0.015268} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.595516] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-710119fc-9290-4223-839f-d98cf1c80eac {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.603276] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 904.603276] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52642377-31d8-f5c3-73f5-352dd849fae6" [ 904.603276] env[62952]: _type = "Task" [ 904.603276] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.614793] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52642377-31d8-f5c3-73f5-352dd849fae6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.687845] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263349, 'name': ReconfigVM_Task, 'duration_secs': 0.576836} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.688142] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 86ec7b3e-34d6-45e8-87b2-c8550f035ed0/86ec7b3e-34d6-45e8-87b2-c8550f035ed0.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.688776] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c4db615-1907-49a7-85cd-2b7091516b1a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.697343] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 904.697343] env[62952]: value = "task-1263351" [ 904.697343] env[62952]: _type = "Task" [ 904.697343] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.707263] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263351, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.772079] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.773321] env[62952]: DEBUG nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 904.775976] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.789s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.777973] env[62952]: INFO nova.compute.claims [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 904.804026] env[62952]: DEBUG oslo_vmware.rw_handles [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529242b2-99e9-5f09-9c4b-de99a85bbbe3/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 904.804715] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f64d03c7-8fff-491d-9b5d-ef05cf0fe40a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.812030] env[62952]: DEBUG oslo_vmware.rw_handles [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529242b2-99e9-5f09-9c4b-de99a85bbbe3/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 904.812210] env[62952]: ERROR oslo_vmware.rw_handles [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529242b2-99e9-5f09-9c4b-de99a85bbbe3/disk-0.vmdk due to incomplete transfer. [ 904.812694] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-44c927d8-709d-48ff-97fe-d0ba3adfb5f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.822643] env[62952]: DEBUG oslo_vmware.rw_handles [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529242b2-99e9-5f09-9c4b-de99a85bbbe3/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 904.822858] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Uploaded image 315e9b97-1fa4-4f98-9bd7-cd708e6717cd to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 904.825087] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 904.825579] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-afde43d0-a7f3-43b7-aa48-8f73a7c1a0b3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.832617] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 904.832617] env[62952]: value = "task-1263352" [ 904.832617] env[62952]: _type = "Task" [ 904.832617] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.842283] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263352, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.904920] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263350, 'name': CreateVM_Task, 'duration_secs': 0.402055} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.905130] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 904.905541] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.905710] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.906165] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 904.906429] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07e56f78-666b-481b-8466-619cb3ce37cd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.912176] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 904.912176] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d4944c-a24e-9f11-9fbe-c79b809a145c" [ 904.912176] env[62952]: _type = "Task" [ 904.912176] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.920976] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d4944c-a24e-9f11-9fbe-c79b809a145c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.114457] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52642377-31d8-f5c3-73f5-352dd849fae6, 'name': SearchDatastore_Task, 'duration_secs': 0.015036} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.114824] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.115168] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 3416eb38-b961-4d48-8b37-8b7b92375dc9/3416eb38-b961-4d48-8b37-8b7b92375dc9.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 905.115656] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b5f2facd-8f70-489a-91a6-400bf731fe04 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.123074] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 905.123074] env[62952]: value = "task-1263353" [ 905.123074] env[62952]: _type = "Task" [ 905.123074] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.131568] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263353, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.207109] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263351, 'name': Rename_Task, 'duration_secs': 0.286221} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.207109] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 905.207317] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a74b2e3a-6437-4035-aa87-b4fc58f29f23 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.214845] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 905.214845] env[62952]: value = "task-1263354" [ 905.214845] env[62952]: _type = "Task" [ 905.214845] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.224498] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263354, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.284532] env[62952]: DEBUG nova.compute.utils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 905.288272] env[62952]: DEBUG nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 905.288272] env[62952]: DEBUG nova.network.neutron [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 905.344881] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263352, 'name': Destroy_Task} progress is 33%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.346771] env[62952]: DEBUG nova.policy [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8fdf784c61c44acc87d6fede9b4b150f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eac1e69a14e94fb89ca59c8d0b5edfaf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 905.423408] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d4944c-a24e-9f11-9fbe-c79b809a145c, 'name': SearchDatastore_Task, 'duration_secs': 0.016149} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.423743] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.424085] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 905.424310] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.424461] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.424655] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.424945] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-24db3b9c-8cd7-41ce-b540-14ba68455376 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.434115] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 905.434307] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 905.435411] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-600cfa58-63ea-4ef8-9b84-f2b63e7530df {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.440886] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 905.440886] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524337db-ebd7-c25f-3454-3cc6c1099495" [ 905.440886] env[62952]: _type = "Task" [ 905.440886] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.449258] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524337db-ebd7-c25f-3454-3cc6c1099495, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.634507] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263353, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.726329] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263354, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.769057] env[62952]: DEBUG nova.compute.manager [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 905.770191] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab111a5-e717-487c-a841-df2ad06d692e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.789206] env[62952]: DEBUG nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 905.851966] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263352, 'name': Destroy_Task, 'duration_secs': 0.55484} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.852358] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Destroyed the VM [ 905.852642] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 905.852956] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0bb81cda-d04e-430f-86a3-f540b02a33ae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.869806] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 905.869806] env[62952]: value = "task-1263355" [ 905.869806] env[62952]: _type = "Task" [ 905.869806] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.879632] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263355, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.956769] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524337db-ebd7-c25f-3454-3cc6c1099495, 'name': SearchDatastore_Task, 'duration_secs': 0.02278} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.960501] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fa50862-da82-4499-a7f2-3b4d1c2fe67f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.967321] env[62952]: DEBUG nova.network.neutron [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Successfully created port: 2580ff6b-2a37-4562-9f37-62a853d44f71 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 905.971299] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 905.971299] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a47ed5-637e-ba16-88bf-8acff422cc74" [ 905.971299] env[62952]: _type = "Task" [ 905.971299] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.984633] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a47ed5-637e-ba16-88bf-8acff422cc74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.137903] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263353, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.885237} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.141192] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 3416eb38-b961-4d48-8b37-8b7b92375dc9/3416eb38-b961-4d48-8b37-8b7b92375dc9.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 906.141513] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 906.142076] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de0b1180-773c-4b24-a04f-a1efca5b394a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.152303] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 906.152303] env[62952]: value = "task-1263356" [ 906.152303] env[62952]: _type = "Task" [ 906.152303] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.167086] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263356, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.216757] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a2ee4bc-3afd-4706-b606-8f09f9e288bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.232736] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263354, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.233894] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-969ce878-9a7f-458b-9231-dc93ff9aa943 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.273358] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a0346fd-7d43-4882-9206-e5560eeeff33 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.282894] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d70eb7-031f-47b5-bc3a-d8ead28be0b9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.288032] env[62952]: INFO nova.compute.manager [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] instance snapshotting [ 906.290967] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5b8ff6-bf4b-456b-87e2-142f253c36a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.305712] env[62952]: DEBUG nova.compute.provider_tree [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.326672] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b38ad3-d95e-4c9e-9d8f-bc41187ac54b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.380320] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263355, 'name': RemoveSnapshot_Task, 'duration_secs': 0.435334} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.380596] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 906.380865] env[62952]: DEBUG nova.compute.manager [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.381717] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b771547-bbef-43aa-8b8b-a4baf78d067c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.482179] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a47ed5-637e-ba16-88bf-8acff422cc74, 'name': SearchDatastore_Task, 'duration_secs': 0.017809} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.482477] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.482741] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 47b57348-40d1-4735-9f5a-a7288f8f12ee/47b57348-40d1-4735-9f5a-a7288f8f12ee.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 906.483013] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-24e47908-fa56-4e68-bbd1-710b6dfc2641 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.490951] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 906.490951] env[62952]: value = "task-1263357" [ 906.490951] env[62952]: _type = "Task" [ 906.490951] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.499489] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263357, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.665417] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263356, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085661} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.665731] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 906.666544] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8fa79a-89fa-45e1-9a12-a96c80e53340 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.690058] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 3416eb38-b961-4d48-8b37-8b7b92375dc9/3416eb38-b961-4d48-8b37-8b7b92375dc9.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 906.690793] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9aeca477-f528-47cb-acea-60ec1b34f91f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.711088] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 906.711088] env[62952]: value = "task-1263358" [ 906.711088] env[62952]: _type = "Task" [ 906.711088] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.719291] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263358, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.728658] env[62952]: DEBUG oslo_vmware.api [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263354, 'name': PowerOnVM_Task, 'duration_secs': 1.027121} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.728936] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 906.729144] env[62952]: INFO nova.compute.manager [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Took 6.97 seconds to spawn the instance on the hypervisor. [ 906.729600] env[62952]: DEBUG nova.compute.manager [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.730421] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74143a63-ed1e-405b-baf4-034aa7a08dd6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.810065] env[62952]: DEBUG nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 906.812723] env[62952]: DEBUG nova.scheduler.client.report [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 906.839196] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 906.841505] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 906.841733] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 906.841894] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 906.842097] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 906.842249] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 906.842398] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 906.842605] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 906.842769] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 906.842940] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 906.843116] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 906.843294] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 906.843910] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4de24ec1-1332-4f54-8181-3c5eb2df5a65 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.846892] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ffd6040-4e4e-46e9-9409-df8e1d55ecf2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.856242] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a890d2d5-2ed7-4a47-b1fe-7b4d3cbd5d03 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.860582] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 906.860582] env[62952]: value = "task-1263359" [ 906.860582] env[62952]: _type = "Task" [ 906.860582] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.878795] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263359, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.896998] env[62952]: INFO nova.compute.manager [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Shelve offloading [ 906.898748] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 906.899049] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1ea1279f-ac52-4c88-809e-c285cf7d2cf8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.907585] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 906.907585] env[62952]: value = "task-1263360" [ 906.907585] env[62952]: _type = "Task" [ 906.907585] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.915944] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263360, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.001801] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263357, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.223139] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263358, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.249630] env[62952]: INFO nova.compute.manager [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Took 34.57 seconds to build instance. [ 907.318564] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.319217] env[62952]: DEBUG nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 907.321965] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.238s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.324042] env[62952]: INFO nova.compute.claims [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 907.378602] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263359, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.418756] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 907.419132] env[62952]: DEBUG nova.compute.manager [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 907.419735] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a129ac9-5bd9-48da-9d63-c5ab415278a3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.426161] env[62952]: DEBUG oslo_concurrency.lockutils [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.426352] env[62952]: DEBUG oslo_concurrency.lockutils [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.426525] env[62952]: DEBUG nova.network.neutron [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 907.503189] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263357, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.950531} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.503772] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 47b57348-40d1-4735-9f5a-a7288f8f12ee/47b57348-40d1-4735-9f5a-a7288f8f12ee.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 907.503772] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 907.503941] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4fcdb8a1-2304-4809-816b-d81ecbdc7c56 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.511750] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 907.511750] env[62952]: value = "task-1263361" [ 907.511750] env[62952]: _type = "Task" [ 907.511750] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.519781] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263361, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.613810] env[62952]: DEBUG nova.compute.manager [req-62857669-99f7-43c2-ba98-489bdafc6514 req-107e297f-b8a7-4fba-9e8b-bfa416c7eb11 service nova] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Received event network-vif-plugged-2580ff6b-2a37-4562-9f37-62a853d44f71 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 907.614129] env[62952]: DEBUG oslo_concurrency.lockutils [req-62857669-99f7-43c2-ba98-489bdafc6514 req-107e297f-b8a7-4fba-9e8b-bfa416c7eb11 service nova] Acquiring lock "e5e7b2b5-34cf-4d37-87ff-1b57264a3516-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.614358] env[62952]: DEBUG oslo_concurrency.lockutils [req-62857669-99f7-43c2-ba98-489bdafc6514 req-107e297f-b8a7-4fba-9e8b-bfa416c7eb11 service nova] Lock "e5e7b2b5-34cf-4d37-87ff-1b57264a3516-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.614533] env[62952]: DEBUG oslo_concurrency.lockutils [req-62857669-99f7-43c2-ba98-489bdafc6514 req-107e297f-b8a7-4fba-9e8b-bfa416c7eb11 service nova] Lock "e5e7b2b5-34cf-4d37-87ff-1b57264a3516-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.614781] env[62952]: DEBUG nova.compute.manager [req-62857669-99f7-43c2-ba98-489bdafc6514 req-107e297f-b8a7-4fba-9e8b-bfa416c7eb11 service nova] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] No waiting events found dispatching network-vif-plugged-2580ff6b-2a37-4562-9f37-62a853d44f71 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 907.614978] env[62952]: WARNING nova.compute.manager [req-62857669-99f7-43c2-ba98-489bdafc6514 req-107e297f-b8a7-4fba-9e8b-bfa416c7eb11 service nova] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Received unexpected event network-vif-plugged-2580ff6b-2a37-4562-9f37-62a853d44f71 for instance with vm_state building and task_state spawning. [ 907.722245] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263358, 'name': ReconfigVM_Task, 'duration_secs': 0.610797} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.722552] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 3416eb38-b961-4d48-8b37-8b7b92375dc9/3416eb38-b961-4d48-8b37-8b7b92375dc9.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 907.723199] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-572e0207-9c97-40ea-a447-17ad13afaba2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.729641] env[62952]: DEBUG nova.network.neutron [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Successfully updated port: 2580ff6b-2a37-4562-9f37-62a853d44f71 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 907.732408] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 907.732408] env[62952]: value = "task-1263362" [ 907.732408] env[62952]: _type = "Task" [ 907.732408] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.741685] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263362, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.754056] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f0c6363-5be3-4db3-9367-57efbe75d0be tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "86ec7b3e-34d6-45e8-87b2-c8550f035ed0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.469s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.829369] env[62952]: DEBUG nova.compute.utils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 907.833274] env[62952]: DEBUG nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 907.833274] env[62952]: DEBUG nova.network.neutron [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 907.874367] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263359, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.875716] env[62952]: DEBUG nova.policy [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8fdf784c61c44acc87d6fede9b4b150f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eac1e69a14e94fb89ca59c8d0b5edfaf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 908.022409] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263361, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.10387} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.022409] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 908.022868] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9dfb215-3adc-44d0-90be-dfaf4733e521 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.043421] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 47b57348-40d1-4735-9f5a-a7288f8f12ee/47b57348-40d1-4735-9f5a-a7288f8f12ee.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 908.045791] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e81c873-b9b2-4b78-b737-eb9aa19623cd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.066879] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 908.066879] env[62952]: value = "task-1263363" [ 908.066879] env[62952]: _type = "Task" [ 908.066879] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.075771] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263363, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.186969] env[62952]: DEBUG nova.network.neutron [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Successfully created port: 36b66914-e58d-4d5f-98c6-e37217eab93c {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 908.201186] env[62952]: DEBUG nova.network.neutron [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Updating instance_info_cache with network_info: [{"id": "3a705583-6448-4020-9894-5f2056e2e4e7", "address": "fa:16:3e:86:6e:fa", "network": {"id": "247a88e9-e8e8-46af-ac41-6777f0782fe0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1297422061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d93803f89ca64e66b8271e7b0bb48983", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a705583-64", "ovs_interfaceid": "3a705583-6448-4020-9894-5f2056e2e4e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.236224] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "refresh_cache-e5e7b2b5-34cf-4d37-87ff-1b57264a3516" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.236385] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquired lock "refresh_cache-e5e7b2b5-34cf-4d37-87ff-1b57264a3516" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.236534] env[62952]: DEBUG nova.network.neutron [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 908.243217] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263362, 'name': Rename_Task, 'duration_secs': 0.434905} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.243794] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 908.244529] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5b93a444-440c-4a6e-b3fb-495f2e44d4c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.252085] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 908.252085] env[62952]: value = "task-1263364" [ 908.252085] env[62952]: _type = "Task" [ 908.252085] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.260903] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263364, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.334155] env[62952]: DEBUG nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 908.376291] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263359, 'name': CreateSnapshot_Task, 'duration_secs': 1.373043} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.380041] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 908.381140] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-189b857c-d3d8-4712-9dcb-823560368af0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.583129] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263363, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.670762] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd79fc68-841d-456e-893d-4d5f37ee2d52 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.680718] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-321ede02-0684-4f1c-9869-f5be5e024ddf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.715591] env[62952]: DEBUG oslo_concurrency.lockutils [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Releasing lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.718173] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436889a6-2544-4e5a-b810-ba51f8ae1ddc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.726884] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dfc692f-9bf1-4682-8a07-55e719a96ec6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.748097] env[62952]: DEBUG nova.compute.provider_tree [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.763268] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263364, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.792226] env[62952]: DEBUG nova.network.neutron [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 908.912102] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 908.912910] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a893c3bb-5636-40ac-b194-031da1b4f946 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.924023] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 908.924023] env[62952]: value = "task-1263365" [ 908.924023] env[62952]: _type = "Task" [ 908.924023] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.932509] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263365, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.055392] env[62952]: DEBUG nova.network.neutron [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Updating instance_info_cache with network_info: [{"id": "2580ff6b-2a37-4562-9f37-62a853d44f71", "address": "fa:16:3e:df:41:26", "network": {"id": "618b6e15-3a3e-4cd5-a48d-7e040179b32d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-798457818-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eac1e69a14e94fb89ca59c8d0b5edfaf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2580ff6b-2a", "ovs_interfaceid": "2580ff6b-2a37-4562-9f37-62a853d44f71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.078418] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263363, 'name': ReconfigVM_Task, 'duration_secs': 0.680861} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.078698] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 47b57348-40d1-4735-9f5a-a7288f8f12ee/47b57348-40d1-4735-9f5a-a7288f8f12ee.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 909.079333] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-882df267-d3f1-475b-a6f5-47b965f19d51 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.087116] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 909.087116] env[62952]: value = "task-1263366" [ 909.087116] env[62952]: _type = "Task" [ 909.087116] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.096580] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263366, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.167591] env[62952]: DEBUG nova.compute.manager [req-ecb6d739-791b-4d86-a26e-b1380968b83d req-7c965d82-4309-4b36-a21d-e20d8c63d7e0 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Received event network-vif-unplugged-3a705583-6448-4020-9894-5f2056e2e4e7 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 909.167833] env[62952]: DEBUG oslo_concurrency.lockutils [req-ecb6d739-791b-4d86-a26e-b1380968b83d req-7c965d82-4309-4b36-a21d-e20d8c63d7e0 service nova] Acquiring lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.170620] env[62952]: DEBUG oslo_concurrency.lockutils [req-ecb6d739-791b-4d86-a26e-b1380968b83d req-7c965d82-4309-4b36-a21d-e20d8c63d7e0 service nova] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.170620] env[62952]: DEBUG oslo_concurrency.lockutils [req-ecb6d739-791b-4d86-a26e-b1380968b83d req-7c965d82-4309-4b36-a21d-e20d8c63d7e0 service nova] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.170620] env[62952]: DEBUG nova.compute.manager [req-ecb6d739-791b-4d86-a26e-b1380968b83d req-7c965d82-4309-4b36-a21d-e20d8c63d7e0 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] No waiting events found dispatching network-vif-unplugged-3a705583-6448-4020-9894-5f2056e2e4e7 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 909.170620] env[62952]: WARNING nova.compute.manager [req-ecb6d739-791b-4d86-a26e-b1380968b83d req-7c965d82-4309-4b36-a21d-e20d8c63d7e0 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Received unexpected event network-vif-unplugged-3a705583-6448-4020-9894-5f2056e2e4e7 for instance with vm_state shelved and task_state shelving_offloading. [ 909.251807] env[62952]: DEBUG nova.scheduler.client.report [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.265143] env[62952]: DEBUG oslo_vmware.api [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263364, 'name': PowerOnVM_Task, 'duration_secs': 0.733379} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.265462] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 909.265686] env[62952]: DEBUG nova.compute.manager [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.267243] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d7de07-e93e-4fba-9a01-f08913e41482 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.345323] env[62952]: DEBUG nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 909.368232] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 909.368493] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 909.368649] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 909.368845] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 909.368996] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 909.369267] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 909.369349] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 909.369509] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 909.369678] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 909.369843] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 909.370031] env[62952]: DEBUG nova.virt.hardware [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 909.370928] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cba261e-ec3b-4ca2-9f94-3123c05cc30a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.380217] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aae46f0-d287-4302-9b69-7adda264c796 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.436672] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263365, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.440183] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 909.441365] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102741d5-f603-4fac-8e9a-e08d9e760b97 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.450055] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 909.450340] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1dc568a9-c51f-473d-80d8-f60205ddc0ed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.548750] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 909.548979] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 909.549322] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Deleting the datastore file [datastore1] 9cc7bc2f-ff22-43cf-be79-ff8c6426115b {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 909.549668] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d48e0b08-215b-4f79-8aa7-be7bc0a71302 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.557356] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 909.557356] env[62952]: value = "task-1263368" [ 909.557356] env[62952]: _type = "Task" [ 909.557356] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.559039] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Releasing lock "refresh_cache-e5e7b2b5-34cf-4d37-87ff-1b57264a3516" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.559039] env[62952]: DEBUG nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Instance network_info: |[{"id": "2580ff6b-2a37-4562-9f37-62a853d44f71", "address": "fa:16:3e:df:41:26", "network": {"id": "618b6e15-3a3e-4cd5-a48d-7e040179b32d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-798457818-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eac1e69a14e94fb89ca59c8d0b5edfaf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2580ff6b-2a", "ovs_interfaceid": "2580ff6b-2a37-4562-9f37-62a853d44f71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 909.562210] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:41:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8800a981-a89e-42e4-8be9-cace419ba9cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2580ff6b-2a37-4562-9f37-62a853d44f71', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 909.569855] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Creating folder: Project (eac1e69a14e94fb89ca59c8d0b5edfaf). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 909.570861] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a3658b9e-6af9-46be-9cad-a269b8216c23 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.578951] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263368, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.582468] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Created folder: Project (eac1e69a14e94fb89ca59c8d0b5edfaf) in parent group-v271811. [ 909.582670] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Creating folder: Instances. Parent ref: group-v271896. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 909.582923] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-73746d1e-1868-4029-8559-1047e8819d6b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.592698] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Created folder: Instances in parent group-v271896. [ 909.592999] env[62952]: DEBUG oslo.service.loopingcall [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 909.593636] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 909.593990] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8465dc34-45b9-4626-9b83-f1af0a5c47fa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.611549] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263366, 'name': Rename_Task, 'duration_secs': 0.20723} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.612293] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 909.612594] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0abc0800-6665-485a-91c5-998fb2ed38ba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.618121] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 909.618121] env[62952]: value = "task-1263371" [ 909.618121] env[62952]: _type = "Task" [ 909.618121] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.619409] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 909.619409] env[62952]: value = "task-1263372" [ 909.619409] env[62952]: _type = "Task" [ 909.619409] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.630335] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263371, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.735910] env[62952]: DEBUG nova.compute.manager [req-cd3bd01e-27d8-47ee-9830-5286283a7d34 req-53c6a173-4d16-4fd0-b278-5acf87587b33 service nova] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Received event network-changed-2580ff6b-2a37-4562-9f37-62a853d44f71 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 909.736189] env[62952]: DEBUG nova.compute.manager [req-cd3bd01e-27d8-47ee-9830-5286283a7d34 req-53c6a173-4d16-4fd0-b278-5acf87587b33 service nova] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Refreshing instance network info cache due to event network-changed-2580ff6b-2a37-4562-9f37-62a853d44f71. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 909.736425] env[62952]: DEBUG oslo_concurrency.lockutils [req-cd3bd01e-27d8-47ee-9830-5286283a7d34 req-53c6a173-4d16-4fd0-b278-5acf87587b33 service nova] Acquiring lock "refresh_cache-e5e7b2b5-34cf-4d37-87ff-1b57264a3516" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.736570] env[62952]: DEBUG oslo_concurrency.lockutils [req-cd3bd01e-27d8-47ee-9830-5286283a7d34 req-53c6a173-4d16-4fd0-b278-5acf87587b33 service nova] Acquired lock "refresh_cache-e5e7b2b5-34cf-4d37-87ff-1b57264a3516" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.736730] env[62952]: DEBUG nova.network.neutron [req-cd3bd01e-27d8-47ee-9830-5286283a7d34 req-53c6a173-4d16-4fd0-b278-5acf87587b33 service nova] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Refreshing network info cache for port 2580ff6b-2a37-4562-9f37-62a853d44f71 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 909.760796] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.761383] env[62952]: DEBUG nova.compute.manager [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 909.764186] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.440s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.764280] env[62952]: DEBUG nova.objects.instance [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Lazy-loading 'resources' on Instance uuid 2b11ad9b-63d8-40da-8f68-3d8260c63bc7 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 909.788072] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.937087] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263365, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.069504] env[62952]: DEBUG oslo_vmware.api [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263368, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173626} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.069811] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.070018] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 910.070208] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 910.111875] env[62952]: INFO nova.scheduler.client.report [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Deleted allocations for instance 9cc7bc2f-ff22-43cf-be79-ff8c6426115b [ 910.118418] env[62952]: DEBUG nova.network.neutron [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Successfully updated port: 36b66914-e58d-4d5f-98c6-e37217eab93c {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 910.136142] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263372, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.136424] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263371, 'name': CreateVM_Task, 'duration_secs': 0.372619} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.136586] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 910.138034] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.138034] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.138163] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 910.139102] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7741913d-b0d2-48bc-bf43-3b1d82452550 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.145806] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 910.145806] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a9ccc9-2549-576a-8878-675a6e15603a" [ 910.145806] env[62952]: _type = "Task" [ 910.145806] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.155457] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a9ccc9-2549-576a-8878-675a6e15603a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.270354] env[62952]: DEBUG nova.compute.utils [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 910.279968] env[62952]: DEBUG nova.compute.manager [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 910.279968] env[62952]: DEBUG nova.network.neutron [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 910.332312] env[62952]: DEBUG nova.policy [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9f6f0133f4a4ac287d4d8d3f4992754', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1487c06883e444a589d9b688155dc00e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 910.436612] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263365, 'name': CloneVM_Task, 'duration_secs': 1.469438} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.439891] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Created linked-clone VM from snapshot [ 910.441158] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bdf8182-9664-4215-8318-d1ea4d0d7a1d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.450967] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Uploading image ab3330a5-e542-4678-beec-181220f16b72 {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 910.487388] env[62952]: DEBUG oslo_vmware.rw_handles [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 910.487388] env[62952]: value = "vm-271895" [ 910.487388] env[62952]: _type = "VirtualMachine" [ 910.487388] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 910.488127] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3f3852c6-7040-4bbd-a7aa-f75084c96430 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.497423] env[62952]: DEBUG oslo_vmware.rw_handles [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lease: (returnval){ [ 910.497423] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ce1bea-30e1-1c4e-cae8-8da0f08385b9" [ 910.497423] env[62952]: _type = "HttpNfcLease" [ 910.497423] env[62952]: } obtained for exporting VM: (result){ [ 910.497423] env[62952]: value = "vm-271895" [ 910.497423] env[62952]: _type = "VirtualMachine" [ 910.497423] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 910.497899] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the lease: (returnval){ [ 910.497899] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ce1bea-30e1-1c4e-cae8-8da0f08385b9" [ 910.497899] env[62952]: _type = "HttpNfcLease" [ 910.497899] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 910.508606] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 910.508606] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ce1bea-30e1-1c4e-cae8-8da0f08385b9" [ 910.508606] env[62952]: _type = "HttpNfcLease" [ 910.508606] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 910.545322] env[62952]: DEBUG nova.network.neutron [req-cd3bd01e-27d8-47ee-9830-5286283a7d34 req-53c6a173-4d16-4fd0-b278-5acf87587b33 service nova] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Updated VIF entry in instance network info cache for port 2580ff6b-2a37-4562-9f37-62a853d44f71. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 910.545720] env[62952]: DEBUG nova.network.neutron [req-cd3bd01e-27d8-47ee-9830-5286283a7d34 req-53c6a173-4d16-4fd0-b278-5acf87587b33 service nova] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Updating instance_info_cache with network_info: [{"id": "2580ff6b-2a37-4562-9f37-62a853d44f71", "address": "fa:16:3e:df:41:26", "network": {"id": "618b6e15-3a3e-4cd5-a48d-7e040179b32d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-798457818-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eac1e69a14e94fb89ca59c8d0b5edfaf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2580ff6b-2a", "ovs_interfaceid": "2580ff6b-2a37-4562-9f37-62a853d44f71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.593631] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b6e622-9da3-4afe-bebf-839a20160ee8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.603716] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b348c3-8cc8-420e-ad4b-a33e4629fe75 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.641249] env[62952]: DEBUG oslo_concurrency.lockutils [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.642068] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "refresh_cache-051f9279-533a-4545-b9e2-272bf40956f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.642208] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquired lock "refresh_cache-051f9279-533a-4545-b9e2-272bf40956f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.642357] env[62952]: DEBUG nova.network.neutron [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 910.644386] env[62952]: DEBUG nova.network.neutron [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Successfully created port: c6818fec-4bb8-4a9d-9106-6ef1f220347b {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 910.649433] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7a24143-2ae0-46e5-ac6d-0d2d6cb8abb8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.661212] env[62952]: DEBUG oslo_vmware.api [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263372, 'name': PowerOnVM_Task, 'duration_secs': 0.663118} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.666490] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 910.666638] env[62952]: INFO nova.compute.manager [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Took 6.36 seconds to spawn the instance on the hypervisor. [ 910.666819] env[62952]: DEBUG nova.compute.manager [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 910.667462] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a9ccc9-2549-576a-8878-675a6e15603a, 'name': SearchDatastore_Task, 'duration_secs': 0.011946} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.668138] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca9b9389-05d9-44a6-94f0-5b8f171bc6a0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.671466] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654aeb05-235c-4910-8e7b-eed797796b30 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.675769] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.676111] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 910.676411] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.676598] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.676808] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 910.677378] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6095a319-3c2f-410f-9dda-c295c8d7bf5c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.690018] env[62952]: DEBUG nova.compute.provider_tree [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.693508] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 910.693696] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 910.697386] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b1e6d38-8b5f-4eff-8b72-838c0ab78def {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.700394] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "3416eb38-b961-4d48-8b37-8b7b92375dc9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.701103] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "3416eb38-b961-4d48-8b37-8b7b92375dc9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.701103] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "3416eb38-b961-4d48-8b37-8b7b92375dc9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.701103] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "3416eb38-b961-4d48-8b37-8b7b92375dc9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.701256] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "3416eb38-b961-4d48-8b37-8b7b92375dc9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.703378] env[62952]: INFO nova.compute.manager [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Terminating instance [ 910.705137] env[62952]: DEBUG nova.compute.manager [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 910.705331] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 910.706683] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c5330f-a9f7-404d-a13d-280711163b5c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.710700] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 910.710700] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52654fdb-9f8f-887f-64ac-648154b91310" [ 910.710700] env[62952]: _type = "Task" [ 910.710700] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.716568] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 910.717152] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ca0f3c3-1a28-4a46-8941-27242549cbcf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.722023] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52654fdb-9f8f-887f-64ac-648154b91310, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.725085] env[62952]: DEBUG oslo_vmware.api [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 910.725085] env[62952]: value = "task-1263374" [ 910.725085] env[62952]: _type = "Task" [ 910.725085] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.734023] env[62952]: DEBUG oslo_vmware.api [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263374, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.776861] env[62952]: DEBUG nova.compute.manager [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 911.011684] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 911.011684] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ce1bea-30e1-1c4e-cae8-8da0f08385b9" [ 911.011684] env[62952]: _type = "HttpNfcLease" [ 911.011684] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 911.011684] env[62952]: DEBUG oslo_vmware.rw_handles [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 911.011684] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ce1bea-30e1-1c4e-cae8-8da0f08385b9" [ 911.011684] env[62952]: _type = "HttpNfcLease" [ 911.011684] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 911.011684] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70cf9a5f-bf0a-4914-83be-7f92ea267cd9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.019703] env[62952]: DEBUG oslo_vmware.rw_handles [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52eb6070-e31e-eaed-f472-89621b0037ce/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 911.019885] env[62952]: DEBUG oslo_vmware.rw_handles [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52eb6070-e31e-eaed-f472-89621b0037ce/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 911.087365] env[62952]: DEBUG oslo_concurrency.lockutils [req-cd3bd01e-27d8-47ee-9830-5286283a7d34 req-53c6a173-4d16-4fd0-b278-5acf87587b33 service nova] Releasing lock "refresh_cache-e5e7b2b5-34cf-4d37-87ff-1b57264a3516" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.188822] env[62952]: DEBUG nova.network.neutron [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 911.199642] env[62952]: DEBUG nova.scheduler.client.report [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 911.215559] env[62952]: INFO nova.compute.manager [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Took 26.61 seconds to build instance. [ 911.220648] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-042d4023-931a-4454-930d-31f14724b14b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.235991] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52654fdb-9f8f-887f-64ac-648154b91310, 'name': SearchDatastore_Task, 'duration_secs': 0.019847} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.239422] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae3c5497-b12f-4763-969f-2a161fdab5b5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.246349] env[62952]: DEBUG oslo_vmware.api [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263374, 'name': PowerOffVM_Task, 'duration_secs': 0.239388} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.249627] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 911.251306] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 911.251306] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ba2a0a66-b5e6-426b-a300-f3f117a3e90f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.254066] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 911.254066] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5255688a-8d77-ba7f-fc20-cffd07b9f965" [ 911.254066] env[62952]: _type = "Task" [ 911.254066] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.271650] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5255688a-8d77-ba7f-fc20-cffd07b9f965, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.326018] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 911.326018] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 911.326018] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleting the datastore file [datastore1] 3416eb38-b961-4d48-8b37-8b7b92375dc9 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 911.326018] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c583d1e-f808-431b-b187-7d438a24198c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.334934] env[62952]: DEBUG oslo_vmware.api [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 911.334934] env[62952]: value = "task-1263376" [ 911.334934] env[62952]: _type = "Task" [ 911.334934] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.346808] env[62952]: DEBUG oslo_vmware.api [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263376, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.388855] env[62952]: DEBUG nova.compute.manager [req-029b8159-e8b8-4b33-b060-49f438ca1417 req-ba8ba120-42e4-4cb3-b254-bb9fdd692b36 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Received event network-changed-3a705583-6448-4020-9894-5f2056e2e4e7 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 911.388855] env[62952]: DEBUG nova.compute.manager [req-029b8159-e8b8-4b33-b060-49f438ca1417 req-ba8ba120-42e4-4cb3-b254-bb9fdd692b36 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Refreshing instance network info cache due to event network-changed-3a705583-6448-4020-9894-5f2056e2e4e7. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 911.388855] env[62952]: DEBUG oslo_concurrency.lockutils [req-029b8159-e8b8-4b33-b060-49f438ca1417 req-ba8ba120-42e4-4cb3-b254-bb9fdd692b36 service nova] Acquiring lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.388855] env[62952]: DEBUG oslo_concurrency.lockutils [req-029b8159-e8b8-4b33-b060-49f438ca1417 req-ba8ba120-42e4-4cb3-b254-bb9fdd692b36 service nova] Acquired lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.388855] env[62952]: DEBUG nova.network.neutron [req-029b8159-e8b8-4b33-b060-49f438ca1417 req-ba8ba120-42e4-4cb3-b254-bb9fdd692b36 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Refreshing network info cache for port 3a705583-6448-4020-9894-5f2056e2e4e7 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 911.412374] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "e17b3d83-7cea-4a65-9457-7e57a6e7641a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.412626] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "e17b3d83-7cea-4a65-9457-7e57a6e7641a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.426852] env[62952]: DEBUG nova.network.neutron [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Updating instance_info_cache with network_info: [{"id": "36b66914-e58d-4d5f-98c6-e37217eab93c", "address": "fa:16:3e:8b:cd:7b", "network": {"id": "618b6e15-3a3e-4cd5-a48d-7e040179b32d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-798457818-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eac1e69a14e94fb89ca59c8d0b5edfaf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap36b66914-e5", "ovs_interfaceid": "36b66914-e58d-4d5f-98c6-e37217eab93c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.705429] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.941s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.708015] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.854s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.708491] env[62952]: DEBUG nova.objects.instance [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lazy-loading 'resources' on Instance uuid dd700ee9-0957-4210-be50-e4c8ac3ca456 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 911.716783] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95602d14-4204-41dc-a523-73aa7b0a1667 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "47b57348-40d1-4735-9f5a-a7288f8f12ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.189s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.730868] env[62952]: INFO nova.scheduler.client.report [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Deleted allocations for instance 2b11ad9b-63d8-40da-8f68-3d8260c63bc7 [ 911.765632] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5255688a-8d77-ba7f-fc20-cffd07b9f965, 'name': SearchDatastore_Task, 'duration_secs': 0.019368} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.766526] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.766526] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] e5e7b2b5-34cf-4d37-87ff-1b57264a3516/e5e7b2b5-34cf-4d37-87ff-1b57264a3516.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 911.767051] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-64505da9-d117-4515-ab1b-aec0c9a17f3b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.772687] env[62952]: DEBUG nova.compute.manager [req-32c3e6b2-5e6a-4e97-bb21-c5abc5bcc0ef req-31b4d515-5071-4a0e-ad23-bbbdbfe98211 service nova] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Received event network-vif-plugged-36b66914-e58d-4d5f-98c6-e37217eab93c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 911.773566] env[62952]: DEBUG oslo_concurrency.lockutils [req-32c3e6b2-5e6a-4e97-bb21-c5abc5bcc0ef req-31b4d515-5071-4a0e-ad23-bbbdbfe98211 service nova] Acquiring lock "051f9279-533a-4545-b9e2-272bf40956f6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.773930] env[62952]: DEBUG oslo_concurrency.lockutils [req-32c3e6b2-5e6a-4e97-bb21-c5abc5bcc0ef req-31b4d515-5071-4a0e-ad23-bbbdbfe98211 service nova] Lock "051f9279-533a-4545-b9e2-272bf40956f6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.774659] env[62952]: DEBUG oslo_concurrency.lockutils [req-32c3e6b2-5e6a-4e97-bb21-c5abc5bcc0ef req-31b4d515-5071-4a0e-ad23-bbbdbfe98211 service nova] Lock "051f9279-533a-4545-b9e2-272bf40956f6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.774659] env[62952]: DEBUG nova.compute.manager [req-32c3e6b2-5e6a-4e97-bb21-c5abc5bcc0ef req-31b4d515-5071-4a0e-ad23-bbbdbfe98211 service nova] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] No waiting events found dispatching network-vif-plugged-36b66914-e58d-4d5f-98c6-e37217eab93c {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.774949] env[62952]: WARNING nova.compute.manager [req-32c3e6b2-5e6a-4e97-bb21-c5abc5bcc0ef req-31b4d515-5071-4a0e-ad23-bbbdbfe98211 service nova] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Received unexpected event network-vif-plugged-36b66914-e58d-4d5f-98c6-e37217eab93c for instance with vm_state building and task_state spawning. [ 911.775256] env[62952]: DEBUG nova.compute.manager [req-32c3e6b2-5e6a-4e97-bb21-c5abc5bcc0ef req-31b4d515-5071-4a0e-ad23-bbbdbfe98211 service nova] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Received event network-changed-36b66914-e58d-4d5f-98c6-e37217eab93c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 911.775447] env[62952]: DEBUG nova.compute.manager [req-32c3e6b2-5e6a-4e97-bb21-c5abc5bcc0ef req-31b4d515-5071-4a0e-ad23-bbbdbfe98211 service nova] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Refreshing instance network info cache due to event network-changed-36b66914-e58d-4d5f-98c6-e37217eab93c. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 911.775926] env[62952]: DEBUG oslo_concurrency.lockutils [req-32c3e6b2-5e6a-4e97-bb21-c5abc5bcc0ef req-31b4d515-5071-4a0e-ad23-bbbdbfe98211 service nova] Acquiring lock "refresh_cache-051f9279-533a-4545-b9e2-272bf40956f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.777658] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 911.777658] env[62952]: value = "task-1263377" [ 911.777658] env[62952]: _type = "Task" [ 911.777658] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.792025] env[62952]: DEBUG nova.compute.manager [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 911.793049] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263377, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.819420] env[62952]: DEBUG nova.virt.hardware [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='73800e59839626efcbc5174d98cf09dd',container_format='bare',created_at=2024-10-10T21:34:56Z,direct_url=,disk_format='vmdk',id=8518bf0a-b7da-4699-a0d0-32c8ad4958d8,min_disk=1,min_ram=0,name='tempest-test-snap-2084754997',owner='1487c06883e444a589d9b688155dc00e',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-10T21:35:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 911.819848] env[62952]: DEBUG nova.virt.hardware [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 911.820264] env[62952]: DEBUG nova.virt.hardware [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 911.820692] env[62952]: DEBUG nova.virt.hardware [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 911.820804] env[62952]: DEBUG nova.virt.hardware [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 911.820958] env[62952]: DEBUG nova.virt.hardware [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 911.821307] env[62952]: DEBUG nova.virt.hardware [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 911.821499] env[62952]: DEBUG nova.virt.hardware [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 911.821742] env[62952]: DEBUG nova.virt.hardware [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 911.821975] env[62952]: DEBUG nova.virt.hardware [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 911.822256] env[62952]: DEBUG nova.virt.hardware [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 911.823331] env[62952]: INFO nova.compute.manager [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Rebuilding instance [ 911.826211] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e3897a-2852-4824-b656-881754903e69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.840213] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a98b40f8-ac00-4fb0-926d-587e0c358ee8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.866393] env[62952]: DEBUG oslo_vmware.api [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263376, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.274821} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.866845] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.867125] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 911.867315] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 911.867480] env[62952]: INFO nova.compute.manager [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Took 1.16 seconds to destroy the instance on the hypervisor. [ 911.867761] env[62952]: DEBUG oslo.service.loopingcall [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.867983] env[62952]: DEBUG nova.compute.manager [-] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 911.868140] env[62952]: DEBUG nova.network.neutron [-] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 911.887359] env[62952]: DEBUG nova.compute.manager [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.888274] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715eae18-f852-4e7a-ac24-231edf8acc25 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.915191] env[62952]: DEBUG nova.compute.manager [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 911.930190] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Releasing lock "refresh_cache-051f9279-533a-4545-b9e2-272bf40956f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.930544] env[62952]: DEBUG nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Instance network_info: |[{"id": "36b66914-e58d-4d5f-98c6-e37217eab93c", "address": "fa:16:3e:8b:cd:7b", "network": {"id": "618b6e15-3a3e-4cd5-a48d-7e040179b32d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-798457818-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eac1e69a14e94fb89ca59c8d0b5edfaf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap36b66914-e5", "ovs_interfaceid": "36b66914-e58d-4d5f-98c6-e37217eab93c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 911.930876] env[62952]: DEBUG oslo_concurrency.lockutils [req-32c3e6b2-5e6a-4e97-bb21-c5abc5bcc0ef req-31b4d515-5071-4a0e-ad23-bbbdbfe98211 service nova] Acquired lock "refresh_cache-051f9279-533a-4545-b9e2-272bf40956f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.931075] env[62952]: DEBUG nova.network.neutron [req-32c3e6b2-5e6a-4e97-bb21-c5abc5bcc0ef req-31b4d515-5071-4a0e-ad23-bbbdbfe98211 service nova] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Refreshing network info cache for port 36b66914-e58d-4d5f-98c6-e37217eab93c {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 911.933221] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8b:cd:7b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8800a981-a89e-42e4-8be9-cace419ba9cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '36b66914-e58d-4d5f-98c6-e37217eab93c', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 911.944158] env[62952]: DEBUG oslo.service.loopingcall [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.951943] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 911.952838] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6675528a-b660-4d6d-b6f0-25de502031ea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.978387] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 911.978387] env[62952]: value = "task-1263378" [ 911.978387] env[62952]: _type = "Task" [ 911.978387] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.987717] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263378, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.247886] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b463e65-2595-475a-a4b1-a2c2af1645b7 tempest-VolumesAdminNegativeTest-1864914686 tempest-VolumesAdminNegativeTest-1864914686-project-member] Lock "2b11ad9b-63d8-40da-8f68-3d8260c63bc7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.077s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.297739] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263377, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.381645] env[62952]: DEBUG nova.network.neutron [req-029b8159-e8b8-4b33-b060-49f438ca1417 req-ba8ba120-42e4-4cb3-b254-bb9fdd692b36 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Updated VIF entry in instance network info cache for port 3a705583-6448-4020-9894-5f2056e2e4e7. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 912.382939] env[62952]: DEBUG nova.network.neutron [req-029b8159-e8b8-4b33-b060-49f438ca1417 req-ba8ba120-42e4-4cb3-b254-bb9fdd692b36 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Updating instance_info_cache with network_info: [{"id": "3a705583-6448-4020-9894-5f2056e2e4e7", "address": "fa:16:3e:86:6e:fa", "network": {"id": "247a88e9-e8e8-46af-ac41-6777f0782fe0", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1297422061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d93803f89ca64e66b8271e7b0bb48983", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap3a705583-64", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.405086] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 912.407226] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3c828192-e2ba-4cd3-8071-cbacccb5e74b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.414088] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 912.414088] env[62952]: value = "task-1263379" [ 912.414088] env[62952]: _type = "Task" [ 912.414088] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.426102] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263379, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.442912] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.464699] env[62952]: DEBUG nova.network.neutron [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Successfully updated port: c6818fec-4bb8-4a9d-9106-6ef1f220347b {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 912.495913] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263378, 'name': CreateVM_Task, 'duration_secs': 0.493042} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.498565] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 912.499474] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.499640] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.499979] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 912.500326] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd4f6745-fea1-4f41-8534-a5fa29785a79 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.507027] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 912.507027] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52abe7eb-9398-a598-1c02-cdb754932c2f" [ 912.507027] env[62952]: _type = "Task" [ 912.507027] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.518784] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52abe7eb-9398-a598-1c02-cdb754932c2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.638545] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f36022c4-070f-4368-9d45-d526dc2f085c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.647405] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4070b3c5-0b7a-4210-ae17-256f3654aa54 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.680041] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94746f11-6edb-410c-ac79-5d0604d85384 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.685325] env[62952]: DEBUG nova.network.neutron [req-32c3e6b2-5e6a-4e97-bb21-c5abc5bcc0ef req-31b4d515-5071-4a0e-ad23-bbbdbfe98211 service nova] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Updated VIF entry in instance network info cache for port 36b66914-e58d-4d5f-98c6-e37217eab93c. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 912.685687] env[62952]: DEBUG nova.network.neutron [req-32c3e6b2-5e6a-4e97-bb21-c5abc5bcc0ef req-31b4d515-5071-4a0e-ad23-bbbdbfe98211 service nova] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Updating instance_info_cache with network_info: [{"id": "36b66914-e58d-4d5f-98c6-e37217eab93c", "address": "fa:16:3e:8b:cd:7b", "network": {"id": "618b6e15-3a3e-4cd5-a48d-7e040179b32d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-798457818-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eac1e69a14e94fb89ca59c8d0b5edfaf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap36b66914-e5", "ovs_interfaceid": "36b66914-e58d-4d5f-98c6-e37217eab93c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.690158] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-876d89dc-942c-4edb-8006-e4afaaa06b12 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.707170] env[62952]: DEBUG nova.compute.provider_tree [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.725095] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.791472] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263377, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.591456} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.792691] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] e5e7b2b5-34cf-4d37-87ff-1b57264a3516/e5e7b2b5-34cf-4d37-87ff-1b57264a3516.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 912.793032] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 912.793400] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-733f9567-a70c-4976-a93f-d151677ea700 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.804049] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 912.804049] env[62952]: value = "task-1263380" [ 912.804049] env[62952]: _type = "Task" [ 912.804049] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.815310] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263380, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.885574] env[62952]: DEBUG oslo_concurrency.lockutils [req-029b8159-e8b8-4b33-b060-49f438ca1417 req-ba8ba120-42e4-4cb3-b254-bb9fdd692b36 service nova] Releasing lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.925560] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263379, 'name': PowerOffVM_Task, 'duration_secs': 0.195048} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.926022] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 912.926379] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 912.927266] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c676d192-038c-4856-9e0f-c17b716eb42d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.937029] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 912.937029] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-054c3e95-9c9a-407d-82d2-288d9b24981a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.965132] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 912.965281] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 912.965525] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Deleting the datastore file [datastore1] 47b57348-40d1-4735-9f5a-a7288f8f12ee {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 912.966097] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c95c769-6fbc-401e-83fa-e0c80dc88f7d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.971290] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "refresh_cache-7a73cb49-3e93-4fcc-b613-3808a8d5c64a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.971290] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquired lock "refresh_cache-7a73cb49-3e93-4fcc-b613-3808a8d5c64a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.971415] env[62952]: DEBUG nova.network.neutron [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 912.975192] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 912.975192] env[62952]: value = "task-1263382" [ 912.975192] env[62952]: _type = "Task" [ 912.975192] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.983670] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263382, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.019689] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52abe7eb-9398-a598-1c02-cdb754932c2f, 'name': SearchDatastore_Task, 'duration_secs': 0.015597} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.020134] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.020436] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 913.020718] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.020922] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.021291] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 913.021619] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8dbb142d-a638-4494-b282-4de4ee25079b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.045200] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 913.045520] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 913.046284] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bc54184-883c-4682-a55f-a38bf32e2734 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.053548] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 913.053548] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5237e1bb-4635-9df3-ba8f-a7516ab31dd8" [ 913.053548] env[62952]: _type = "Task" [ 913.053548] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.063196] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5237e1bb-4635-9df3-ba8f-a7516ab31dd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.189780] env[62952]: DEBUG oslo_concurrency.lockutils [req-32c3e6b2-5e6a-4e97-bb21-c5abc5bcc0ef req-31b4d515-5071-4a0e-ad23-bbbdbfe98211 service nova] Releasing lock "refresh_cache-051f9279-533a-4545-b9e2-272bf40956f6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.210913] env[62952]: DEBUG nova.scheduler.client.report [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 913.276044] env[62952]: DEBUG nova.network.neutron [-] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.315849] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263380, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077794} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.319733] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 913.320968] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd7f380-111c-4640-b87f-4f92f033c7a2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.348514] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] e5e7b2b5-34cf-4d37-87ff-1b57264a3516/e5e7b2b5-34cf-4d37-87ff-1b57264a3516.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.348893] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b2f0b3d-e4da-48a6-9960-8d9dd20ed763 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.372158] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 913.372158] env[62952]: value = "task-1263383" [ 913.372158] env[62952]: _type = "Task" [ 913.372158] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.383339] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263383, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.486832] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263382, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.216313} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.487113] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 913.487310] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 913.487491] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 913.561742] env[62952]: DEBUG nova.network.neutron [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 913.569726] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5237e1bb-4635-9df3-ba8f-a7516ab31dd8, 'name': SearchDatastore_Task, 'duration_secs': 0.016617} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.570712] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0180418-0fd4-4d8a-84bd-56e21e2e6248 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.577625] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 913.577625] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b52666-272a-6dec-4bab-f5cb418fa146" [ 913.577625] env[62952]: _type = "Task" [ 913.577625] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.587907] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b52666-272a-6dec-4bab-f5cb418fa146, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.665098] env[62952]: DEBUG nova.compute.manager [req-e33aba81-f7cc-423e-944a-4ce36d7e1424 req-16a1fc90-a3de-4e6d-8d82-ca8c890891e9 service nova] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Received event network-vif-plugged-c6818fec-4bb8-4a9d-9106-6ef1f220347b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 913.665344] env[62952]: DEBUG oslo_concurrency.lockutils [req-e33aba81-f7cc-423e-944a-4ce36d7e1424 req-16a1fc90-a3de-4e6d-8d82-ca8c890891e9 service nova] Acquiring lock "7a73cb49-3e93-4fcc-b613-3808a8d5c64a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.665552] env[62952]: DEBUG oslo_concurrency.lockutils [req-e33aba81-f7cc-423e-944a-4ce36d7e1424 req-16a1fc90-a3de-4e6d-8d82-ca8c890891e9 service nova] Lock "7a73cb49-3e93-4fcc-b613-3808a8d5c64a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.665721] env[62952]: DEBUG oslo_concurrency.lockutils [req-e33aba81-f7cc-423e-944a-4ce36d7e1424 req-16a1fc90-a3de-4e6d-8d82-ca8c890891e9 service nova] Lock "7a73cb49-3e93-4fcc-b613-3808a8d5c64a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.665888] env[62952]: DEBUG nova.compute.manager [req-e33aba81-f7cc-423e-944a-4ce36d7e1424 req-16a1fc90-a3de-4e6d-8d82-ca8c890891e9 service nova] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] No waiting events found dispatching network-vif-plugged-c6818fec-4bb8-4a9d-9106-6ef1f220347b {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 913.666114] env[62952]: WARNING nova.compute.manager [req-e33aba81-f7cc-423e-944a-4ce36d7e1424 req-16a1fc90-a3de-4e6d-8d82-ca8c890891e9 service nova] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Received unexpected event network-vif-plugged-c6818fec-4bb8-4a9d-9106-6ef1f220347b for instance with vm_state building and task_state spawning. [ 913.666298] env[62952]: DEBUG nova.compute.manager [req-e33aba81-f7cc-423e-944a-4ce36d7e1424 req-16a1fc90-a3de-4e6d-8d82-ca8c890891e9 service nova] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Received event network-changed-c6818fec-4bb8-4a9d-9106-6ef1f220347b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 913.666463] env[62952]: DEBUG nova.compute.manager [req-e33aba81-f7cc-423e-944a-4ce36d7e1424 req-16a1fc90-a3de-4e6d-8d82-ca8c890891e9 service nova] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Refreshing instance network info cache due to event network-changed-c6818fec-4bb8-4a9d-9106-6ef1f220347b. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 913.666618] env[62952]: DEBUG oslo_concurrency.lockutils [req-e33aba81-f7cc-423e-944a-4ce36d7e1424 req-16a1fc90-a3de-4e6d-8d82-ca8c890891e9 service nova] Acquiring lock "refresh_cache-7a73cb49-3e93-4fcc-b613-3808a8d5c64a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.715122] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.007s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.717736] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.088s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.717973] env[62952]: DEBUG nova.objects.instance [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Lazy-loading 'resources' on Instance uuid 219cfb14-ca73-45d3-86e8-ed4642d7d480 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.742045] env[62952]: INFO nova.scheduler.client.report [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleted allocations for instance dd700ee9-0957-4210-be50-e4c8ac3ca456 [ 913.776676] env[62952]: INFO nova.compute.manager [-] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Took 1.91 seconds to deallocate network for instance. [ 913.883995] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263383, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.936549] env[62952]: DEBUG nova.network.neutron [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Updating instance_info_cache with network_info: [{"id": "c6818fec-4bb8-4a9d-9106-6ef1f220347b", "address": "fa:16:3e:12:85:d8", "network": {"id": "2c5146eb-dcff-470d-984f-b96a3c91bb3e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1950178234-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1487c06883e444a589d9b688155dc00e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6818fec-4b", "ovs_interfaceid": "c6818fec-4bb8-4a9d-9106-6ef1f220347b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.955330] env[62952]: DEBUG nova.compute.manager [req-4e0f66b2-55e4-4228-bab4-3daf4cfbe623 req-1e8ab2b4-ba45-4018-ac44-68163c4a0181 service nova] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Received event network-vif-deleted-d4ed51c5-2a91-40ed-8ce9-62f6109ab3b8 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 914.090423] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b52666-272a-6dec-4bab-f5cb418fa146, 'name': SearchDatastore_Task, 'duration_secs': 0.018772} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.090735] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.090989] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 051f9279-533a-4545-b9e2-272bf40956f6/051f9279-533a-4545-b9e2-272bf40956f6.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 914.091350] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bf9f4142-1648-4f64-98f8-2aa887bbab36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.102304] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 914.102304] env[62952]: value = "task-1263384" [ 914.102304] env[62952]: _type = "Task" [ 914.102304] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.113157] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263384, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.253013] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e705f822-be91-48fc-890a-4b17836f32e1 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "dd700ee9-0957-4210-be50-e4c8ac3ca456" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.897s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.287140] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.391331] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263383, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.441656] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Releasing lock "refresh_cache-7a73cb49-3e93-4fcc-b613-3808a8d5c64a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.442177] env[62952]: DEBUG nova.compute.manager [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Instance network_info: |[{"id": "c6818fec-4bb8-4a9d-9106-6ef1f220347b", "address": "fa:16:3e:12:85:d8", "network": {"id": "2c5146eb-dcff-470d-984f-b96a3c91bb3e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1950178234-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1487c06883e444a589d9b688155dc00e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6818fec-4b", "ovs_interfaceid": "c6818fec-4bb8-4a9d-9106-6ef1f220347b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 914.442531] env[62952]: DEBUG oslo_concurrency.lockutils [req-e33aba81-f7cc-423e-944a-4ce36d7e1424 req-16a1fc90-a3de-4e6d-8d82-ca8c890891e9 service nova] Acquired lock "refresh_cache-7a73cb49-3e93-4fcc-b613-3808a8d5c64a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.442715] env[62952]: DEBUG nova.network.neutron [req-e33aba81-f7cc-423e-944a-4ce36d7e1424 req-16a1fc90-a3de-4e6d-8d82-ca8c890891e9 service nova] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Refreshing network info cache for port c6818fec-4bb8-4a9d-9106-6ef1f220347b {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 914.444265] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:12:85:d8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e445fb59-822c-4d7d-943b-c8e3bbaca62e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c6818fec-4bb8-4a9d-9106-6ef1f220347b', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 914.453895] env[62952]: DEBUG oslo.service.loopingcall [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.458913] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 914.459207] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-991ac67e-aeb2-4cc1-b1dd-6e5374a08ae4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.488463] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.488463] env[62952]: value = "task-1263385" [ 914.488463] env[62952]: _type = "Task" [ 914.488463] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.502282] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263385, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.540502] env[62952]: DEBUG nova.virt.hardware [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 914.540502] env[62952]: DEBUG nova.virt.hardware [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 914.540733] env[62952]: DEBUG nova.virt.hardware [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 914.541451] env[62952]: DEBUG nova.virt.hardware [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 914.541451] env[62952]: DEBUG nova.virt.hardware [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 914.541732] env[62952]: DEBUG nova.virt.hardware [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 914.542286] env[62952]: DEBUG nova.virt.hardware [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 914.542603] env[62952]: DEBUG nova.virt.hardware [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 914.543140] env[62952]: DEBUG nova.virt.hardware [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 914.543391] env[62952]: DEBUG nova.virt.hardware [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 914.543721] env[62952]: DEBUG nova.virt.hardware [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 914.545469] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-663cd782-b81f-40a0-a391-55533dad07fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.560058] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40daa0b6-a8ed-4c70-b826-c073c97f646a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.579370] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 914.586347] env[62952]: DEBUG oslo.service.loopingcall [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.593462] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 914.594066] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b2bc71f-e3fd-45a5-936c-def7886d01f3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.626619] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263384, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.628705] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.628705] env[62952]: value = "task-1263386" [ 914.628705] env[62952]: _type = "Task" [ 914.628705] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.633480] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12b05bb-145f-445e-b15f-93168ea196d3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.653210] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f021232-1dec-4e21-a556-95bd86ba531a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.658594] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263386, 'name': CreateVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.701777] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-077bb292-8f2b-478a-a658-edf82df4c176 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.715358] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6939c75d-ec25-4cf3-baca-fb4aafaedf42 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.728969] env[62952]: DEBUG nova.compute.provider_tree [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.885971] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263383, 'name': ReconfigVM_Task, 'duration_secs': 1.291958} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.886495] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Reconfigured VM instance instance-00000045 to attach disk [datastore1] e5e7b2b5-34cf-4d37-87ff-1b57264a3516/e5e7b2b5-34cf-4d37-87ff-1b57264a3516.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.887197] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-31ed288c-661f-4005-a268-af2aaed530b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.896885] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 914.896885] env[62952]: value = "task-1263387" [ 914.896885] env[62952]: _type = "Task" [ 914.896885] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.913157] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263387, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.946716] env[62952]: DEBUG nova.network.neutron [req-e33aba81-f7cc-423e-944a-4ce36d7e1424 req-16a1fc90-a3de-4e6d-8d82-ca8c890891e9 service nova] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Updated VIF entry in instance network info cache for port c6818fec-4bb8-4a9d-9106-6ef1f220347b. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 914.947184] env[62952]: DEBUG nova.network.neutron [req-e33aba81-f7cc-423e-944a-4ce36d7e1424 req-16a1fc90-a3de-4e6d-8d82-ca8c890891e9 service nova] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Updating instance_info_cache with network_info: [{"id": "c6818fec-4bb8-4a9d-9106-6ef1f220347b", "address": "fa:16:3e:12:85:d8", "network": {"id": "2c5146eb-dcff-470d-984f-b96a3c91bb3e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1950178234-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1487c06883e444a589d9b688155dc00e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6818fec-4b", "ovs_interfaceid": "c6818fec-4bb8-4a9d-9106-6ef1f220347b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.001288] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263385, 'name': CreateVM_Task, 'duration_secs': 0.500274} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.001478] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 915.002737] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8518bf0a-b7da-4699-a0d0-32c8ad4958d8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.002940] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8518bf0a-b7da-4699-a0d0-32c8ad4958d8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.003386] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8518bf0a-b7da-4699-a0d0-32c8ad4958d8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.003885] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-499e5de9-8785-4f32-b18d-1a1d9774f52f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.010285] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 915.010285] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525f054a-ebe1-79d4-27bc-767fa6f783ae" [ 915.010285] env[62952]: _type = "Task" [ 915.010285] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.024533] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525f054a-ebe1-79d4-27bc-767fa6f783ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.124801] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263384, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.579629} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.125150] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 051f9279-533a-4545-b9e2-272bf40956f6/051f9279-533a-4545-b9e2-272bf40956f6.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 915.125437] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.125743] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb77c8b9-3c96-4b90-9d41-dee41a568e19 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.133610] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 915.133610] env[62952]: value = "task-1263388" [ 915.133610] env[62952]: _type = "Task" [ 915.133610] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.146845] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263388, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.150105] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263386, 'name': CreateVM_Task, 'duration_secs': 0.362599} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.150289] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 915.150723] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.150896] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.151259] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.151536] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24e3e173-3e65-4080-860a-555901c47492 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.157646] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 915.157646] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523ee5db-e954-6f45-8b4d-b5f85b87f31e" [ 915.157646] env[62952]: _type = "Task" [ 915.157646] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.169653] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523ee5db-e954-6f45-8b4d-b5f85b87f31e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.232861] env[62952]: DEBUG nova.scheduler.client.report [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 915.409731] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263387, 'name': Rename_Task, 'duration_secs': 0.199675} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.409731] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 915.409731] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-10c8d65e-6cdc-41ef-bad9-6ccccd3d2aa3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.417800] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 915.417800] env[62952]: value = "task-1263389" [ 915.417800] env[62952]: _type = "Task" [ 915.417800] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.427060] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263389, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.452881] env[62952]: DEBUG oslo_concurrency.lockutils [req-e33aba81-f7cc-423e-944a-4ce36d7e1424 req-16a1fc90-a3de-4e6d-8d82-ca8c890891e9 service nova] Releasing lock "refresh_cache-7a73cb49-3e93-4fcc-b613-3808a8d5c64a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.522108] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8518bf0a-b7da-4699-a0d0-32c8ad4958d8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.522503] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Processing image 8518bf0a-b7da-4699-a0d0-32c8ad4958d8 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 915.522685] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8518bf0a-b7da-4699-a0d0-32c8ad4958d8/8518bf0a-b7da-4699-a0d0-32c8ad4958d8.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.522761] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8518bf0a-b7da-4699-a0d0-32c8ad4958d8/8518bf0a-b7da-4699-a0d0-32c8ad4958d8.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.522940] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 915.523220] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4227bb38-4041-4b44-b36b-5d2eaadfb8d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.532133] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 915.532367] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 915.533069] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-beff445d-a726-45d1-a28d-43e6b44bcba2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.539252] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 915.539252] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b3bea9-6803-7c67-6d59-c463ed3ba5a9" [ 915.539252] env[62952]: _type = "Task" [ 915.539252] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.547731] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b3bea9-6803-7c67-6d59-c463ed3ba5a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.647119] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263388, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071912} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.647419] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 915.648164] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87afd3e9-0df1-4d42-8b22-449c864504be {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.676244] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 051f9279-533a-4545-b9e2-272bf40956f6/051f9279-533a-4545-b9e2-272bf40956f6.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 915.679517] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e728e993-dc23-4fdd-b568-4402eaf8370b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.700903] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523ee5db-e954-6f45-8b4d-b5f85b87f31e, 'name': SearchDatastore_Task, 'duration_secs': 0.018817} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.702309] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.702666] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 915.702925] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.703091] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.703277] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 915.703607] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 915.703607] env[62952]: value = "task-1263390" [ 915.703607] env[62952]: _type = "Task" [ 915.703607] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.703802] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a547830-f76b-4e8f-bc24-768dc11af2fe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.715587] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263390, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.727012] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 915.727154] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 915.727864] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9df8bf6-6cab-4002-97b5-c8a7f5c18bd9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.734212] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 915.734212] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52eda74f-e3e3-4fa6-91da-064a33cf3c22" [ 915.734212] env[62952]: _type = "Task" [ 915.734212] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.738481] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.021s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.744272] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.154s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.746096] env[62952]: INFO nova.compute.claims [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.749483] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52eda74f-e3e3-4fa6-91da-064a33cf3c22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.758524] env[62952]: INFO nova.scheduler.client.report [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Deleted allocations for instance 219cfb14-ca73-45d3-86e8-ed4642d7d480 [ 915.930485] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263389, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.960555] env[62952]: DEBUG oslo_concurrency.lockutils [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "20ea323b-8961-4981-8045-f9cb85815d99" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.960555] env[62952]: DEBUG oslo_concurrency.lockutils [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "20ea323b-8961-4981-8045-f9cb85815d99" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.051422] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Preparing fetch location {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 916.051568] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Fetch image to [datastore1] OSTACK_IMG_a7a07ff0-706f-4356-b0d5-86359307f98c/OSTACK_IMG_a7a07ff0-706f-4356-b0d5-86359307f98c.vmdk {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 916.051691] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Downloading stream optimized image 8518bf0a-b7da-4699-a0d0-32c8ad4958d8 to [datastore1] OSTACK_IMG_a7a07ff0-706f-4356-b0d5-86359307f98c/OSTACK_IMG_a7a07ff0-706f-4356-b0d5-86359307f98c.vmdk on the data store datastore1 as vApp {{(pid=62952) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 916.051864] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Downloading image file data 8518bf0a-b7da-4699-a0d0-32c8ad4958d8 to the ESX as VM named 'OSTACK_IMG_a7a07ff0-706f-4356-b0d5-86359307f98c' {{(pid=62952) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 916.142556] env[62952]: DEBUG oslo_vmware.rw_handles [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 916.142556] env[62952]: value = "resgroup-9" [ 916.142556] env[62952]: _type = "ResourcePool" [ 916.142556] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 916.143501] env[62952]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-96d99757-a553-4b0d-b90a-aab797d6a1fd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.167128] env[62952]: DEBUG oslo_vmware.rw_handles [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lease: (returnval){ [ 916.167128] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52dc5cce-0ccd-a18e-c161-9d1569b486f1" [ 916.167128] env[62952]: _type = "HttpNfcLease" [ 916.167128] env[62952]: } obtained for vApp import into resource pool (val){ [ 916.167128] env[62952]: value = "resgroup-9" [ 916.167128] env[62952]: _type = "ResourcePool" [ 916.167128] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 916.167128] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the lease: (returnval){ [ 916.167128] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52dc5cce-0ccd-a18e-c161-9d1569b486f1" [ 916.167128] env[62952]: _type = "HttpNfcLease" [ 916.167128] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 916.177560] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 916.177560] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52dc5cce-0ccd-a18e-c161-9d1569b486f1" [ 916.177560] env[62952]: _type = "HttpNfcLease" [ 916.177560] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 916.217460] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263390, 'name': ReconfigVM_Task, 'duration_secs': 0.379609} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.217595] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 051f9279-533a-4545-b9e2-272bf40956f6/051f9279-533a-4545-b9e2-272bf40956f6.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 916.218912] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a9786383-2365-40f6-b59c-0d4963c669d1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.227692] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 916.227692] env[62952]: value = "task-1263392" [ 916.227692] env[62952]: _type = "Task" [ 916.227692] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.243041] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263392, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.250309] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52eda74f-e3e3-4fa6-91da-064a33cf3c22, 'name': SearchDatastore_Task, 'duration_secs': 0.011791} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.253957] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80f1fbcb-008f-4ed3-b6db-770d05478e74 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.261390] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 916.261390] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5234eff4-d8ca-ac83-2611-b6de7ca3529a" [ 916.261390] env[62952]: _type = "Task" [ 916.261390] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.271604] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c935004-7995-42ed-9cdb-e1c53fb1f052 tempest-ServersTestManualDisk-1788027584 tempest-ServersTestManualDisk-1788027584-project-member] Lock "219cfb14-ca73-45d3-86e8-ed4642d7d480" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.214s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.273348] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5234eff4-d8ca-ac83-2611-b6de7ca3529a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.429474] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263389, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.466828] env[62952]: DEBUG nova.compute.manager [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 916.678371] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 916.678371] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52dc5cce-0ccd-a18e-c161-9d1569b486f1" [ 916.678371] env[62952]: _type = "HttpNfcLease" [ 916.678371] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 916.742293] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263392, 'name': Rename_Task, 'duration_secs': 0.229199} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.742616] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 916.742894] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd854831-ad94-4524-bb9e-38eb25a4c5f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.751362] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 916.751362] env[62952]: value = "task-1263393" [ 916.751362] env[62952]: _type = "Task" [ 916.751362] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.763411] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263393, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.777879] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5234eff4-d8ca-ac83-2611-b6de7ca3529a, 'name': SearchDatastore_Task, 'duration_secs': 0.011807} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.778502] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.778793] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 47b57348-40d1-4735-9f5a-a7288f8f12ee/47b57348-40d1-4735-9f5a-a7288f8f12ee.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 916.779146] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-793f18ce-be67-4b67-8a17-437722a3bba4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.788974] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 916.788974] env[62952]: value = "task-1263394" [ 916.788974] env[62952]: _type = "Task" [ 916.788974] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.803473] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263394, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.936265] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263389, 'name': PowerOnVM_Task, 'duration_secs': 1.153088} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.936265] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 916.936767] env[62952]: INFO nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Took 10.13 seconds to spawn the instance on the hypervisor. [ 916.937246] env[62952]: DEBUG nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 916.938743] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bedff46-1dbe-44c9-adf4-46837db3f49c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.003449] env[62952]: DEBUG oslo_concurrency.lockutils [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.184390] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 917.184390] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52dc5cce-0ccd-a18e-c161-9d1569b486f1" [ 917.184390] env[62952]: _type = "HttpNfcLease" [ 917.184390] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 917.184840] env[62952]: DEBUG oslo_vmware.rw_handles [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 917.184840] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52dc5cce-0ccd-a18e-c161-9d1569b486f1" [ 917.184840] env[62952]: _type = "HttpNfcLease" [ 917.184840] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 917.185644] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0b0d396-08cb-415b-b668-031d2132fc8c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.197870] env[62952]: DEBUG oslo_vmware.rw_handles [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ab0241-5e32-332c-0fa6-a630ee1ceb30/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 917.198124] env[62952]: DEBUG oslo_vmware.rw_handles [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ab0241-5e32-332c-0fa6-a630ee1ceb30/disk-0.vmdk. {{(pid=62952) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 917.268878] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09bb354a-3f22-444e-832e-735704eadcd1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.282712] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-162f7114-32b5-4476-8407-f5afd2cadfc7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.285333] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263393, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.290245] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6b6ba94-9e2b-49f1-b8b7-f471d437513c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.304392] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263394, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.332637] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c98876-8db5-4102-b1a2-ee6e50d9aea2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.343059] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dda9ff1-73f0-40f8-a6d8-6bacd57d702a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.362413] env[62952]: DEBUG nova.compute.provider_tree [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.459753] env[62952]: INFO nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Took 30.78 seconds to build instance. [ 917.779032] env[62952]: DEBUG oslo_vmware.api [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263393, 'name': PowerOnVM_Task, 'duration_secs': 0.686291} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.781316] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 917.781316] env[62952]: INFO nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Took 8.43 seconds to spawn the instance on the hypervisor. [ 917.781316] env[62952]: DEBUG nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 917.782039] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d61ba34-06be-4276-84c0-f23e0c046f24 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.810351] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263394, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.566366} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.810720] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 47b57348-40d1-4735-9f5a-a7288f8f12ee/47b57348-40d1-4735-9f5a-a7288f8f12ee.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 917.811071] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 917.811260] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-74ed66f1-044f-4480-878b-0acb7d13f721 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.823154] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 917.823154] env[62952]: value = "task-1263395" [ 917.823154] env[62952]: _type = "Task" [ 917.823154] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.833425] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263395, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.867983] env[62952]: DEBUG nova.scheduler.client.report [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 917.962586] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "e5e7b2b5-34cf-4d37-87ff-1b57264a3516" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.532s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.312074] env[62952]: INFO nova.compute.manager [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Took 28.35 seconds to build instance. [ 918.333774] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263395, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088566} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.336118] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 918.337036] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa2c230f-2362-4f5d-b307-7e20439a3eaa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.366713] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 47b57348-40d1-4735-9f5a-a7288f8f12ee/47b57348-40d1-4735-9f5a-a7288f8f12ee.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 918.368699] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c6b9569-6a04-42a8-a078-e5bff90fdc84 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.392881] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.648s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.393216] env[62952]: DEBUG nova.compute.manager [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 918.397556] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.023s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.402299] env[62952]: INFO nova.compute.claims [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 918.411673] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 918.411673] env[62952]: value = "task-1263396" [ 918.411673] env[62952]: _type = "Task" [ 918.411673] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.429789] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263396, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.503637] env[62952]: DEBUG oslo_vmware.rw_handles [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Completed reading data from the image iterator. {{(pid=62952) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 918.503890] env[62952]: DEBUG oslo_vmware.rw_handles [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ab0241-5e32-332c-0fa6-a630ee1ceb30/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 918.505027] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eda697e8-1adb-4529-85a0-0ec6a203e043 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.515405] env[62952]: DEBUG oslo_vmware.rw_handles [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ab0241-5e32-332c-0fa6-a630ee1ceb30/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 918.515610] env[62952]: DEBUG oslo_vmware.rw_handles [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ab0241-5e32-332c-0fa6-a630ee1ceb30/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 918.516253] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-80208402-ea62-4d13-abad-e817848a2c85 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.814942] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b000cdf2-f53b-4aa1-b362-29a0a0ff206e tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "051f9279-533a-4545-b9e2-272bf40956f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.359s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.875968] env[62952]: DEBUG oslo_vmware.rw_handles [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ab0241-5e32-332c-0fa6-a630ee1ceb30/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 918.876286] env[62952]: INFO nova.virt.vmwareapi.images [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Downloaded image file data 8518bf0a-b7da-4699-a0d0-32c8ad4958d8 [ 918.877417] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffe7b305-7829-4f6d-8b81-3a80c4c96e93 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.908116] env[62952]: DEBUG nova.compute.utils [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.909747] env[62952]: DEBUG nova.compute.manager [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 918.909926] env[62952]: DEBUG nova.network.neutron [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 918.912334] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f50d8d09-a98d-4686-b1c2-4d9d9ffffd32 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.925143] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263396, 'name': ReconfigVM_Task, 'duration_secs': 0.478015} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.927144] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 47b57348-40d1-4735-9f5a-a7288f8f12ee/47b57348-40d1-4735-9f5a-a7288f8f12ee.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 918.927144] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-40150bba-70f5-4e57-a73a-c4fc3b7d254e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.934584] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 918.934584] env[62952]: value = "task-1263398" [ 918.934584] env[62952]: _type = "Task" [ 918.934584] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.946944] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263398, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.948515] env[62952]: INFO nova.virt.vmwareapi.images [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] The imported VM was unregistered [ 918.950273] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Caching image {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 918.950569] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Creating directory with path [datastore1] devstack-image-cache_base/8518bf0a-b7da-4699-a0d0-32c8ad4958d8 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 918.950853] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-af3032db-e82c-438b-825f-370faeec5cc5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.967643] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Created directory with path [datastore1] devstack-image-cache_base/8518bf0a-b7da-4699-a0d0-32c8ad4958d8 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 918.967864] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_a7a07ff0-706f-4356-b0d5-86359307f98c/OSTACK_IMG_a7a07ff0-706f-4356-b0d5-86359307f98c.vmdk to [datastore1] devstack-image-cache_base/8518bf0a-b7da-4699-a0d0-32c8ad4958d8/8518bf0a-b7da-4699-a0d0-32c8ad4958d8.vmdk. {{(pid=62952) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 918.968186] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-e5ad4e32-732a-4007-a9e9-f404e3d1da84 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.984160] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 918.984160] env[62952]: value = "task-1263399" [ 918.984160] env[62952]: _type = "Task" [ 918.984160] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.990119] env[62952]: DEBUG nova.policy [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '867f24e0b3f94882861e1612b1242c66', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ac3c03f232a40ef8e4f353e8323426d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 918.997987] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263399, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.275719] env[62952]: DEBUG oslo_vmware.rw_handles [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52eb6070-e31e-eaed-f472-89621b0037ce/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 919.276925] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ae2823c-7720-4c06-9c9a-31907c6818b4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.286215] env[62952]: DEBUG oslo_concurrency.lockutils [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "e5e7b2b5-34cf-4d37-87ff-1b57264a3516" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.286537] env[62952]: DEBUG oslo_concurrency.lockutils [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "e5e7b2b5-34cf-4d37-87ff-1b57264a3516" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.286710] env[62952]: DEBUG oslo_concurrency.lockutils [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "e5e7b2b5-34cf-4d37-87ff-1b57264a3516-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.286895] env[62952]: DEBUG oslo_concurrency.lockutils [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "e5e7b2b5-34cf-4d37-87ff-1b57264a3516-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.287155] env[62952]: DEBUG oslo_concurrency.lockutils [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "e5e7b2b5-34cf-4d37-87ff-1b57264a3516-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.289350] env[62952]: DEBUG oslo_vmware.rw_handles [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52eb6070-e31e-eaed-f472-89621b0037ce/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 919.289589] env[62952]: ERROR oslo_vmware.rw_handles [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52eb6070-e31e-eaed-f472-89621b0037ce/disk-0.vmdk due to incomplete transfer. [ 919.289778] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a265473f-707e-4dfa-9738-f1a8389877b7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.292343] env[62952]: INFO nova.compute.manager [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Terminating instance [ 919.295177] env[62952]: DEBUG nova.compute.manager [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 919.295392] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 919.296246] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db50d98-1198-4325-b882-6846c41662b9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.301723] env[62952]: DEBUG oslo_vmware.rw_handles [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52eb6070-e31e-eaed-f472-89621b0037ce/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 919.301945] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Uploaded image ab3330a5-e542-4678-beec-181220f16b72 to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 919.304752] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 919.307128] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-76924c89-d703-4bd5-bbf3-b3e32ae5d4a6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.309472] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 919.309775] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31bb0bfc-2434-41b1-8e04-d10b09287c35 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.319120] env[62952]: DEBUG oslo_vmware.api [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 919.319120] env[62952]: value = "task-1263400" [ 919.319120] env[62952]: _type = "Task" [ 919.319120] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.321131] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 919.321131] env[62952]: value = "task-1263401" [ 919.321131] env[62952]: _type = "Task" [ 919.321131] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.339488] env[62952]: DEBUG oslo_vmware.api [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263400, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.344761] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263401, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.415250] env[62952]: DEBUG nova.compute.manager [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 919.451451] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263398, 'name': Rename_Task, 'duration_secs': 0.189472} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.455740] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 919.455740] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2e71ab31-8fe4-40e8-95f8-f413deca2dd3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.468874] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 919.468874] env[62952]: value = "task-1263402" [ 919.468874] env[62952]: _type = "Task" [ 919.468874] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.477347] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "051f9279-533a-4545-b9e2-272bf40956f6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.477628] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "051f9279-533a-4545-b9e2-272bf40956f6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.477945] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "051f9279-533a-4545-b9e2-272bf40956f6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.478038] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "051f9279-533a-4545-b9e2-272bf40956f6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.478255] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "051f9279-533a-4545-b9e2-272bf40956f6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.484262] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263402, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.484262] env[62952]: INFO nova.compute.manager [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Terminating instance [ 919.489607] env[62952]: DEBUG nova.compute.manager [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 919.489995] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 919.491226] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0350a6-4e2b-4266-877d-6e62c9ced820 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.510679] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 919.511466] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263399, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.511845] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-091ebe8b-0325-4d7f-9671-dcee40e9f63b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.521905] env[62952]: DEBUG oslo_vmware.api [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 919.521905] env[62952]: value = "task-1263403" [ 919.521905] env[62952]: _type = "Task" [ 919.521905] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.537650] env[62952]: DEBUG oslo_vmware.api [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263403, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.594860] env[62952]: DEBUG nova.network.neutron [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Successfully created port: a2903c7d-4cae-40ce-8f1a-52d2ea6a9bfd {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 919.842342] env[62952]: DEBUG oslo_vmware.api [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263400, 'name': PowerOffVM_Task, 'duration_secs': 0.258319} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.843421] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 919.843630] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 919.848607] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1b73fb2d-4a8e-4dca-805a-3c9a433d04c8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.856459] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263401, 'name': Destroy_Task} progress is 33%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.883026] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6cc35c-1ca4-441f-88ea-157685e142a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.893665] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7779d66c-3e90-4500-86fa-9c938198f806 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.942195] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-155e89a7-36e2-4e6b-9c45-fa1174242e91 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.949657] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634da08e-fcf9-4970-9bbc-bf9259d72d8b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.972880] env[62952]: DEBUG nova.compute.provider_tree [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.980187] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 919.980187] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 919.980378] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Deleting the datastore file [datastore1] e5e7b2b5-34cf-4d37-87ff-1b57264a3516 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 919.980917] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4c2e2b0f-698b-4484-9264-c45746df888e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.996121] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263402, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.998378] env[62952]: DEBUG oslo_vmware.api [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 919.998378] env[62952]: value = "task-1263405" [ 919.998378] env[62952]: _type = "Task" [ 919.998378] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.008295] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263399, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.014800] env[62952]: DEBUG oslo_vmware.api [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263405, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.036695] env[62952]: DEBUG oslo_vmware.api [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263403, 'name': PowerOffVM_Task, 'duration_secs': 0.268184} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.036999] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 920.037226] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 920.037709] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e3b0a8f-5a5a-4d2c-823c-4ea83a9e66e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.121502] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 920.121502] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 920.121502] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Deleting the datastore file [datastore1] 051f9279-533a-4545-b9e2-272bf40956f6 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 920.121502] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-33b3ccb6-7e68-4d7f-9957-6da1ee63f47c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.130438] env[62952]: DEBUG oslo_vmware.api [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for the task: (returnval){ [ 920.130438] env[62952]: value = "task-1263407" [ 920.130438] env[62952]: _type = "Task" [ 920.130438] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.141537] env[62952]: DEBUG oslo_vmware.api [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263407, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.339053] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263401, 'name': Destroy_Task, 'duration_secs': 0.935414} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.339396] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Destroyed the VM [ 920.339645] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 920.340056] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-333ce926-d5f5-4450-8de3-5f665384d208 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.354602] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 920.354602] env[62952]: value = "task-1263408" [ 920.354602] env[62952]: _type = "Task" [ 920.354602] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.365059] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263408, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.441133] env[62952]: DEBUG nova.compute.manager [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 920.483413] env[62952]: DEBUG nova.virt.hardware [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 920.483609] env[62952]: DEBUG nova.virt.hardware [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 920.483714] env[62952]: DEBUG nova.virt.hardware [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 920.483905] env[62952]: DEBUG nova.virt.hardware [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 920.484197] env[62952]: DEBUG nova.virt.hardware [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 920.484454] env[62952]: DEBUG nova.virt.hardware [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 920.484789] env[62952]: DEBUG nova.virt.hardware [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 920.485916] env[62952]: DEBUG nova.virt.hardware [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 920.485916] env[62952]: DEBUG nova.virt.hardware [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 920.486066] env[62952]: DEBUG nova.virt.hardware [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 920.486300] env[62952]: DEBUG nova.virt.hardware [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 920.487261] env[62952]: DEBUG nova.scheduler.client.report [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 920.491868] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6896f131-e860-4e3b-8a43-5f27ff48ac54 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.499184] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263402, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.515439] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b66d87-5982-4374-8443-746a96383e0b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.528684] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263399, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.529153] env[62952]: DEBUG oslo_vmware.api [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263405, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.643371] env[62952]: DEBUG oslo_vmware.api [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263407, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.865478] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263408, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.985101] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263402, 'name': PowerOnVM_Task} progress is 76%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.996205] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.599s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.996804] env[62952]: DEBUG nova.compute.manager [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 920.999764] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.482s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.999850] env[62952]: DEBUG nova.objects.instance [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Lazy-loading 'resources' on Instance uuid 49947bde-069b-459c-ae19-ae5f090f535b {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.015421] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263399, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.021792] env[62952]: DEBUG oslo_vmware.api [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263405, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.144964] env[62952]: DEBUG oslo_vmware.api [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263407, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.372223] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263408, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.499116] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263402, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.503117] env[62952]: DEBUG nova.compute.utils [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 921.513655] env[62952]: DEBUG nova.compute.manager [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 921.513655] env[62952]: DEBUG nova.network.neutron [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 921.535173] env[62952]: DEBUG oslo_vmware.api [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263405, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.535173] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263399, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.642687] env[62952]: DEBUG oslo_vmware.api [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263407, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.452999} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.645477] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 921.645688] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 921.645868] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 921.646057] env[62952]: INFO nova.compute.manager [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Took 2.16 seconds to destroy the instance on the hypervisor. [ 921.646306] env[62952]: DEBUG oslo.service.loopingcall [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.647444] env[62952]: DEBUG nova.compute.manager [-] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 921.647444] env[62952]: DEBUG nova.network.neutron [-] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 921.685018] env[62952]: DEBUG nova.policy [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6967f8ebdc944dbfb2d19b5d7acdd3e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91fde3ceb94649f698459dfc174bf835', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 921.867359] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263408, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.904079] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e690553-8123-4ffd-8ab3-ac9e555f70f8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.920336] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b35bca66-5d8e-48b0-9357-d7fd7828b03c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.959594] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2099fac1-2f7d-481a-bb5f-b13c17cb6ef8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.969175] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129ddc1e-9991-46b8-97d3-3eda40d549c1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.985864] env[62952]: DEBUG nova.compute.provider_tree [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.006274] env[62952]: DEBUG nova.compute.manager [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 922.016424] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263399, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.653029} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.016842] env[62952]: DEBUG oslo_vmware.api [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263402, 'name': PowerOnVM_Task, 'duration_secs': 2.261596} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.017377] env[62952]: INFO nova.virt.vmwareapi.ds_util [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_a7a07ff0-706f-4356-b0d5-86359307f98c/OSTACK_IMG_a7a07ff0-706f-4356-b0d5-86359307f98c.vmdk to [datastore1] devstack-image-cache_base/8518bf0a-b7da-4699-a0d0-32c8ad4958d8/8518bf0a-b7da-4699-a0d0-32c8ad4958d8.vmdk. [ 922.017573] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Cleaning up location [datastore1] OSTACK_IMG_a7a07ff0-706f-4356-b0d5-86359307f98c {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 922.017740] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_a7a07ff0-706f-4356-b0d5-86359307f98c {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 922.021969] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 922.022098] env[62952]: DEBUG nova.compute.manager [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 922.022376] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-91c551f5-cf80-4e2f-a9a3-f831822d5037 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.025051] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b169e0-47f0-4f57-b7ed-eab851d256db {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.033596] env[62952]: DEBUG oslo_vmware.api [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Task: {'id': task-1263405, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.581634} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.036864] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.037157] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 922.037748] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 922.037961] env[62952]: INFO nova.compute.manager [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Took 2.74 seconds to destroy the instance on the hypervisor. [ 922.038260] env[62952]: DEBUG oslo.service.loopingcall [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.038684] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 922.038684] env[62952]: value = "task-1263409" [ 922.038684] env[62952]: _type = "Task" [ 922.038684] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.041938] env[62952]: DEBUG nova.compute.manager [-] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 922.042200] env[62952]: DEBUG nova.network.neutron [-] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 922.056364] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263409, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.317504] env[62952]: DEBUG nova.network.neutron [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Successfully updated port: a2903c7d-4cae-40ce-8f1a-52d2ea6a9bfd {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 922.332336] env[62952]: DEBUG nova.compute.manager [req-5e27834c-a47f-483a-b671-b5347f9261a2 req-35362f98-9779-450e-a761-991a71f63815 service nova] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Received event network-vif-deleted-36b66914-e58d-4d5f-98c6-e37217eab93c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 922.332559] env[62952]: INFO nova.compute.manager [req-5e27834c-a47f-483a-b671-b5347f9261a2 req-35362f98-9779-450e-a761-991a71f63815 service nova] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Neutron deleted interface 36b66914-e58d-4d5f-98c6-e37217eab93c; detaching it from the instance and deleting it from the info cache [ 922.332738] env[62952]: DEBUG nova.network.neutron [req-5e27834c-a47f-483a-b671-b5347f9261a2 req-35362f98-9779-450e-a761-991a71f63815 service nova] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.368214] env[62952]: DEBUG oslo_vmware.api [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263408, 'name': RemoveSnapshot_Task, 'duration_secs': 1.59875} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.368863] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 922.368863] env[62952]: INFO nova.compute.manager [None req-75513ed0-69eb-46ed-ac07-b6e0137a5c1f tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Took 16.08 seconds to snapshot the instance on the hypervisor. [ 922.459719] env[62952]: DEBUG nova.network.neutron [-] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.488607] env[62952]: DEBUG nova.scheduler.client.report [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.554795] env[62952]: DEBUG nova.compute.manager [req-1c1a139d-7d37-4f8a-b69c-1f45fdedf2b2 req-9e0a6446-d70d-4b43-aa87-06728916218d service nova] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Received event network-vif-plugged-a2903c7d-4cae-40ce-8f1a-52d2ea6a9bfd {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 922.555139] env[62952]: DEBUG oslo_concurrency.lockutils [req-1c1a139d-7d37-4f8a-b69c-1f45fdedf2b2 req-9e0a6446-d70d-4b43-aa87-06728916218d service nova] Acquiring lock "45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.555479] env[62952]: DEBUG oslo_concurrency.lockutils [req-1c1a139d-7d37-4f8a-b69c-1f45fdedf2b2 req-9e0a6446-d70d-4b43-aa87-06728916218d service nova] Lock "45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.555718] env[62952]: DEBUG oslo_concurrency.lockutils [req-1c1a139d-7d37-4f8a-b69c-1f45fdedf2b2 req-9e0a6446-d70d-4b43-aa87-06728916218d service nova] Lock "45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.555937] env[62952]: DEBUG nova.compute.manager [req-1c1a139d-7d37-4f8a-b69c-1f45fdedf2b2 req-9e0a6446-d70d-4b43-aa87-06728916218d service nova] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] No waiting events found dispatching network-vif-plugged-a2903c7d-4cae-40ce-8f1a-52d2ea6a9bfd {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 922.557612] env[62952]: WARNING nova.compute.manager [req-1c1a139d-7d37-4f8a-b69c-1f45fdedf2b2 req-9e0a6446-d70d-4b43-aa87-06728916218d service nova] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Received unexpected event network-vif-plugged-a2903c7d-4cae-40ce-8f1a-52d2ea6a9bfd for instance with vm_state building and task_state spawning. [ 922.557612] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.563763] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263409, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.185578} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.564158] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.564399] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8518bf0a-b7da-4699-a0d0-32c8ad4958d8/8518bf0a-b7da-4699-a0d0-32c8ad4958d8.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.565160] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8518bf0a-b7da-4699-a0d0-32c8ad4958d8/8518bf0a-b7da-4699-a0d0-32c8ad4958d8.vmdk to [datastore1] 7a73cb49-3e93-4fcc-b613-3808a8d5c64a/7a73cb49-3e93-4fcc-b613-3808a8d5c64a.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 922.565160] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1978ce74-2cf4-40ad-b647-ec5e2d7adb29 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.569735] env[62952]: DEBUG nova.network.neutron [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Successfully created port: 6bab534b-286c-4391-aeeb-fbb57655925a {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 922.574369] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 922.574369] env[62952]: value = "task-1263410" [ 922.574369] env[62952]: _type = "Task" [ 922.574369] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.584824] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263410, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.783862] env[62952]: DEBUG nova.network.neutron [-] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.821925] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Acquiring lock "refresh_cache-45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.823104] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Acquired lock "refresh_cache-45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.825300] env[62952]: DEBUG nova.network.neutron [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 922.837412] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1ae6ff05-7960-45e6-ae5c-7ab74635a21e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.850437] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b242fe6f-f2c1-4920-bc57-7ba70d009baf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.885104] env[62952]: DEBUG nova.compute.manager [req-5e27834c-a47f-483a-b671-b5347f9261a2 req-35362f98-9779-450e-a761-991a71f63815 service nova] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Detach interface failed, port_id=36b66914-e58d-4d5f-98c6-e37217eab93c, reason: Instance 051f9279-533a-4545-b9e2-272bf40956f6 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 922.963012] env[62952]: INFO nova.compute.manager [-] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Took 1.31 seconds to deallocate network for instance. [ 922.994464] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.995s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.999664] env[62952]: DEBUG oslo_concurrency.lockutils [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.820s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.999791] env[62952]: DEBUG nova.objects.instance [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Lazy-loading 'resources' on Instance uuid 67b729be-fb66-4ed1-bbea-e62216d460d5 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.024652] env[62952]: DEBUG nova.compute.manager [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 923.030539] env[62952]: INFO nova.scheduler.client.report [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Deleted allocations for instance 49947bde-069b-459c-ae19-ae5f090f535b [ 923.069703] env[62952]: DEBUG nova.virt.hardware [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 923.069853] env[62952]: DEBUG nova.virt.hardware [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 923.075270] env[62952]: DEBUG nova.virt.hardware [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 923.075270] env[62952]: DEBUG nova.virt.hardware [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 923.075270] env[62952]: DEBUG nova.virt.hardware [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 923.075270] env[62952]: DEBUG nova.virt.hardware [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 923.075270] env[62952]: DEBUG nova.virt.hardware [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 923.075960] env[62952]: DEBUG nova.virt.hardware [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 923.075960] env[62952]: DEBUG nova.virt.hardware [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 923.075960] env[62952]: DEBUG nova.virt.hardware [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 923.075960] env[62952]: DEBUG nova.virt.hardware [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 923.075960] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1450ddde-a535-47f9-89fa-ceb6826cab39 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.091068] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5062e724-3b84-46dc-a3c4-0f4173504ce7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.098498] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263410, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.286416] env[62952]: INFO nova.compute.manager [-] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Took 1.24 seconds to deallocate network for instance. [ 923.373654] env[62952]: DEBUG nova.network.neutron [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 923.480097] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.506332] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "47b57348-40d1-4735-9f5a-a7288f8f12ee" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.506332] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "47b57348-40d1-4735-9f5a-a7288f8f12ee" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.506332] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "47b57348-40d1-4735-9f5a-a7288f8f12ee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.506332] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "47b57348-40d1-4735-9f5a-a7288f8f12ee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.506663] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "47b57348-40d1-4735-9f5a-a7288f8f12ee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.510106] env[62952]: INFO nova.compute.manager [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Terminating instance [ 923.513633] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "refresh_cache-47b57348-40d1-4735-9f5a-a7288f8f12ee" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.513932] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquired lock "refresh_cache-47b57348-40d1-4735-9f5a-a7288f8f12ee" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.514261] env[62952]: DEBUG nova.network.neutron [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 923.548832] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4abaf467-d98b-4222-aa83-e8e8f939eaa2 tempest-ServerGroupTestJSON-172482232 tempest-ServerGroupTestJSON-172482232-project-member] Lock "49947bde-069b-459c-ae19-ae5f090f535b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.022s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.596063] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263410, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.597331] env[62952]: DEBUG nova.network.neutron [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Updating instance_info_cache with network_info: [{"id": "a2903c7d-4cae-40ce-8f1a-52d2ea6a9bfd", "address": "fa:16:3e:99:45:a7", "network": {"id": "d1b22c3d-87f5-49fe-ad75-705db9c42844", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1377938858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ac3c03f232a40ef8e4f353e8323426d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2903c7d-4c", "ovs_interfaceid": "a2903c7d-4cae-40ce-8f1a-52d2ea6a9bfd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.801568] env[62952]: DEBUG oslo_concurrency.lockutils [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.937086] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e8a4d2-f199-4fdc-93aa-8f6609ba33be {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.946720] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d6cbe8c-65c9-4e01-8b37-055255aecbea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.982666] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50bc7225-2412-4e35-8d2f-ffd7315fb34e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.993747] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff36f4a7-ee4e-4456-94f4-87e8c9a4bfea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.015450] env[62952]: DEBUG nova.compute.provider_tree [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.045376] env[62952]: DEBUG nova.network.neutron [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 924.092619] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263410, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.100611] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Releasing lock "refresh_cache-45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.101113] env[62952]: DEBUG nova.compute.manager [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Instance network_info: |[{"id": "a2903c7d-4cae-40ce-8f1a-52d2ea6a9bfd", "address": "fa:16:3e:99:45:a7", "network": {"id": "d1b22c3d-87f5-49fe-ad75-705db9c42844", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1377938858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ac3c03f232a40ef8e4f353e8323426d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2903c7d-4c", "ovs_interfaceid": "a2903c7d-4cae-40ce-8f1a-52d2ea6a9bfd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 924.102078] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:45:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '459b8c74-0aa6-42b6-996a-42b1c5d7e5c6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a2903c7d-4cae-40ce-8f1a-52d2ea6a9bfd', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 924.110716] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Creating folder: Project (9ac3c03f232a40ef8e4f353e8323426d). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 924.111077] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-27e1a2ef-765f-40d5-8086-2ec2c3474df4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.126758] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Created folder: Project (9ac3c03f232a40ef8e4f353e8323426d) in parent group-v271811. [ 924.127076] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Creating folder: Instances. Parent ref: group-v271903. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 924.127275] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0d3c9fd6-f7d0-4e7e-b7e2-59f75ce050d4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.140449] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Created folder: Instances in parent group-v271903. [ 924.140872] env[62952]: DEBUG oslo.service.loopingcall [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.141142] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 924.141517] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3cb6e7b8-bfaa-4003-89cf-0c3897fc46a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.162021] env[62952]: DEBUG nova.network.neutron [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.168055] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 924.168055] env[62952]: value = "task-1263413" [ 924.168055] env[62952]: _type = "Task" [ 924.168055] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.180804] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263413, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.521626] env[62952]: DEBUG nova.scheduler.client.report [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 924.591911] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263410, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.600806] env[62952]: DEBUG nova.compute.manager [req-ca87b1d4-abe0-4e48-9aaa-dfdbbf223af6 req-3dd43a32-a16c-436d-968a-f9a0fba29c2d service nova] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Received event network-vif-deleted-2580ff6b-2a37-4562-9f37-62a853d44f71 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 924.666026] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Releasing lock "refresh_cache-47b57348-40d1-4735-9f5a-a7288f8f12ee" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.666026] env[62952]: DEBUG nova.compute.manager [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 924.666026] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 924.666026] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-587cd989-123a-4d00-b2bb-7dfe871e5871 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.681014] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263413, 'name': CreateVM_Task, 'duration_secs': 0.422678} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.683901] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 924.684991] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 924.685900] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.686422] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.687670] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 924.691381] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8156fc5f-4d3d-4ec7-adcd-cba4368fc717 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.692712] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c1c2317-c433-4cb4-9340-5e88b7256a21 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.703267] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Waiting for the task: (returnval){ [ 924.703267] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ea6e2b-017e-23d3-5c9f-294d98799503" [ 924.703267] env[62952]: _type = "Task" [ 924.703267] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.711830] env[62952]: DEBUG oslo_vmware.api [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 924.711830] env[62952]: value = "task-1263414" [ 924.711830] env[62952]: _type = "Task" [ 924.711830] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.723263] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ea6e2b-017e-23d3-5c9f-294d98799503, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.730475] env[62952]: DEBUG oslo_vmware.api [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263414, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.763227] env[62952]: DEBUG nova.compute.manager [req-a4925ab8-71b5-419b-ba5a-f132ec8d3ccc req-448c783f-0d11-445f-b7cb-630b75f2f998 service nova] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Received event network-changed-a2903c7d-4cae-40ce-8f1a-52d2ea6a9bfd {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 924.763227] env[62952]: DEBUG nova.compute.manager [req-a4925ab8-71b5-419b-ba5a-f132ec8d3ccc req-448c783f-0d11-445f-b7cb-630b75f2f998 service nova] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Refreshing instance network info cache due to event network-changed-a2903c7d-4cae-40ce-8f1a-52d2ea6a9bfd. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 924.763713] env[62952]: DEBUG oslo_concurrency.lockutils [req-a4925ab8-71b5-419b-ba5a-f132ec8d3ccc req-448c783f-0d11-445f-b7cb-630b75f2f998 service nova] Acquiring lock "refresh_cache-45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.764035] env[62952]: DEBUG oslo_concurrency.lockutils [req-a4925ab8-71b5-419b-ba5a-f132ec8d3ccc req-448c783f-0d11-445f-b7cb-630b75f2f998 service nova] Acquired lock "refresh_cache-45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.764328] env[62952]: DEBUG nova.network.neutron [req-a4925ab8-71b5-419b-ba5a-f132ec8d3ccc req-448c783f-0d11-445f-b7cb-630b75f2f998 service nova] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Refreshing network info cache for port a2903c7d-4cae-40ce-8f1a-52d2ea6a9bfd {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 924.772199] env[62952]: DEBUG nova.network.neutron [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Successfully updated port: 6bab534b-286c-4391-aeeb-fbb57655925a {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 925.027704] env[62952]: DEBUG oslo_concurrency.lockutils [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.026s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.029198] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.424s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.030232] env[62952]: DEBUG nova.objects.instance [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lazy-loading 'resources' on Instance uuid ba022bea-e357-4541-a573-d1da2eb49b46 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.050304] env[62952]: INFO nova.scheduler.client.report [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Deleted allocations for instance 67b729be-fb66-4ed1-bbea-e62216d460d5 [ 925.092796] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263410, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.216512] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ea6e2b-017e-23d3-5c9f-294d98799503, 'name': SearchDatastore_Task, 'duration_secs': 0.021039} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.221178] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.221575] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 925.222019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.222019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.222165] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 925.222714] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f77f89b-e98b-4975-b350-b72805287b32 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.233219] env[62952]: DEBUG oslo_vmware.api [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263414, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.235844] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 925.236473] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 925.237482] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57c5d2bb-4a4a-4ff1-9760-f121257273c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.246379] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Waiting for the task: (returnval){ [ 925.246379] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525e95cd-b1ec-409a-268c-65ede94da1fc" [ 925.246379] env[62952]: _type = "Task" [ 925.246379] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.258377] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525e95cd-b1ec-409a-268c-65ede94da1fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.276118] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.276118] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.276118] env[62952]: DEBUG nova.network.neutron [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 925.279285] env[62952]: DEBUG nova.compute.manager [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.280210] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a8352f-e87e-4a4b-80a4-935471e0ca0d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.563742] env[62952]: DEBUG oslo_concurrency.lockutils [None req-160d3413-d628-4e1c-a188-199cc7c8d2b2 tempest-ServerRescueTestJSONUnderV235-979677121 tempest-ServerRescueTestJSONUnderV235-979677121-project-member] Lock "67b729be-fb66-4ed1-bbea-e62216d460d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.458s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.597858] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263410, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.859855} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.598156] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8518bf0a-b7da-4699-a0d0-32c8ad4958d8/8518bf0a-b7da-4699-a0d0-32c8ad4958d8.vmdk to [datastore1] 7a73cb49-3e93-4fcc-b613-3808a8d5c64a/7a73cb49-3e93-4fcc-b613-3808a8d5c64a.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 925.599843] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac2b5af9-ee9f-409c-b36d-d9dafe2f5515 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.630713] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 7a73cb49-3e93-4fcc-b613-3808a8d5c64a/7a73cb49-3e93-4fcc-b613-3808a8d5c64a.vmdk or device None with type streamOptimized {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.635622] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a194d6a8-85ea-401d-8fa2-a603e25fea76 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.658868] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 925.658868] env[62952]: value = "task-1263415" [ 925.658868] env[62952]: _type = "Task" [ 925.658868] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.669109] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263415, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.701751] env[62952]: DEBUG nova.network.neutron [req-a4925ab8-71b5-419b-ba5a-f132ec8d3ccc req-448c783f-0d11-445f-b7cb-630b75f2f998 service nova] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Updated VIF entry in instance network info cache for port a2903c7d-4cae-40ce-8f1a-52d2ea6a9bfd. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 925.702116] env[62952]: DEBUG nova.network.neutron [req-a4925ab8-71b5-419b-ba5a-f132ec8d3ccc req-448c783f-0d11-445f-b7cb-630b75f2f998 service nova] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Updating instance_info_cache with network_info: [{"id": "a2903c7d-4cae-40ce-8f1a-52d2ea6a9bfd", "address": "fa:16:3e:99:45:a7", "network": {"id": "d1b22c3d-87f5-49fe-ad75-705db9c42844", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1377938858-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ac3c03f232a40ef8e4f353e8323426d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2903c7d-4c", "ovs_interfaceid": "a2903c7d-4cae-40ce-8f1a-52d2ea6a9bfd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.727646] env[62952]: DEBUG oslo_vmware.api [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263414, 'name': PowerOffVM_Task, 'duration_secs': 0.565254} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.728057] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 925.728283] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 925.728552] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d2dabb20-f618-4e37-ab27-d39d26a8291c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.758760] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525e95cd-b1ec-409a-268c-65ede94da1fc, 'name': SearchDatastore_Task, 'duration_secs': 0.014387} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.761071] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3431606-55dc-427b-8e17-77291baaf127 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.763677] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 925.763897] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 925.764081] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Deleting the datastore file [datastore1] 47b57348-40d1-4735-9f5a-a7288f8f12ee {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.766927] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a482014-2f5b-4659-9f79-45c44b88fde3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.774329] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Waiting for the task: (returnval){ [ 925.774329] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52da330c-a6c7-e990-950b-5f9abd17806e" [ 925.774329] env[62952]: _type = "Task" [ 925.774329] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.774824] env[62952]: DEBUG oslo_vmware.api [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 925.774824] env[62952]: value = "task-1263417" [ 925.774824] env[62952]: _type = "Task" [ 925.774824] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.797404] env[62952]: INFO nova.compute.manager [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] instance snapshotting [ 925.799301] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52da330c-a6c7-e990-950b-5f9abd17806e, 'name': SearchDatastore_Task, 'duration_secs': 0.012809} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.809295] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.809644] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20/45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 925.810460] env[62952]: DEBUG oslo_vmware.api [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263417, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.812499] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b61195-14a8-48b6-8ce0-6194df9879fa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.816080] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b1aa6f9-d909-441a-af5f-214595a5a28f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.843572] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae7937cd-b211-40d7-ba8f-62a48a2e81f4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.847936] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Waiting for the task: (returnval){ [ 925.847936] env[62952]: value = "task-1263418" [ 925.847936] env[62952]: _type = "Task" [ 925.847936] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.861429] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': task-1263418, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.865803] env[62952]: DEBUG nova.network.neutron [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 925.977478] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a27ccc-60e1-45b5-8469-bb4e15f245bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.985838] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b3b6fc4-9010-48b7-b421-e9eecfde6c72 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.035763] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-334c2990-a72b-426e-8f87-b06ea1abfff5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.047112] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1178c13f-d2cd-4d40-8db9-0e3a8f7e0b09 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.065336] env[62952]: DEBUG nova.compute.provider_tree [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.172346] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263415, 'name': ReconfigVM_Task, 'duration_secs': 0.353305} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.172346] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 7a73cb49-3e93-4fcc-b613-3808a8d5c64a/7a73cb49-3e93-4fcc-b613-3808a8d5c64a.vmdk or device None with type streamOptimized {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.172990] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4bc41230-8e6d-4875-a692-1adba630b910 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.182165] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 926.182165] env[62952]: value = "task-1263419" [ 926.182165] env[62952]: _type = "Task" [ 926.182165] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.193286] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263419, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.208356] env[62952]: DEBUG oslo_concurrency.lockutils [req-a4925ab8-71b5-419b-ba5a-f132ec8d3ccc req-448c783f-0d11-445f-b7cb-630b75f2f998 service nova] Releasing lock "refresh_cache-45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.208721] env[62952]: DEBUG nova.compute.manager [req-a4925ab8-71b5-419b-ba5a-f132ec8d3ccc req-448c783f-0d11-445f-b7cb-630b75f2f998 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Received event network-vif-plugged-6bab534b-286c-4391-aeeb-fbb57655925a {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 926.209072] env[62952]: DEBUG oslo_concurrency.lockutils [req-a4925ab8-71b5-419b-ba5a-f132ec8d3ccc req-448c783f-0d11-445f-b7cb-630b75f2f998 service nova] Acquiring lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.209205] env[62952]: DEBUG oslo_concurrency.lockutils [req-a4925ab8-71b5-419b-ba5a-f132ec8d3ccc req-448c783f-0d11-445f-b7cb-630b75f2f998 service nova] Lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.209384] env[62952]: DEBUG oslo_concurrency.lockutils [req-a4925ab8-71b5-419b-ba5a-f132ec8d3ccc req-448c783f-0d11-445f-b7cb-630b75f2f998 service nova] Lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.209597] env[62952]: DEBUG nova.compute.manager [req-a4925ab8-71b5-419b-ba5a-f132ec8d3ccc req-448c783f-0d11-445f-b7cb-630b75f2f998 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] No waiting events found dispatching network-vif-plugged-6bab534b-286c-4391-aeeb-fbb57655925a {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 926.209777] env[62952]: WARNING nova.compute.manager [req-a4925ab8-71b5-419b-ba5a-f132ec8d3ccc req-448c783f-0d11-445f-b7cb-630b75f2f998 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Received unexpected event network-vif-plugged-6bab534b-286c-4391-aeeb-fbb57655925a for instance with vm_state building and task_state spawning. [ 926.292851] env[62952]: DEBUG oslo_vmware.api [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263417, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104272} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.293223] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 926.293489] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 926.293710] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 926.293909] env[62952]: INFO nova.compute.manager [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Took 1.63 seconds to destroy the instance on the hypervisor. [ 926.294212] env[62952]: DEBUG oslo.service.loopingcall [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.294430] env[62952]: DEBUG nova.compute.manager [-] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 926.294567] env[62952]: DEBUG nova.network.neutron [-] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 926.318623] env[62952]: DEBUG nova.network.neutron [-] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 926.358940] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 926.359329] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': task-1263418, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493905} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.359551] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b2c96f32-6cda-4483-a86b-c4f4b71f4c40 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.361541] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20/45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 926.361731] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 926.361979] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-290de371-ee20-4e61-8434-1f5e58545983 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.370021] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Waiting for the task: (returnval){ [ 926.370021] env[62952]: value = "task-1263420" [ 926.370021] env[62952]: _type = "Task" [ 926.370021] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.371424] env[62952]: DEBUG oslo_vmware.api [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 926.371424] env[62952]: value = "task-1263421" [ 926.371424] env[62952]: _type = "Task" [ 926.371424] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.372279] env[62952]: DEBUG nova.network.neutron [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Updating instance_info_cache with network_info: [{"id": "6bab534b-286c-4391-aeeb-fbb57655925a", "address": "fa:16:3e:e2:f4:40", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bab534b-28", "ovs_interfaceid": "6bab534b-286c-4391-aeeb-fbb57655925a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.392516] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': task-1263420, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.392900] env[62952]: DEBUG oslo_vmware.api [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263421, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.568913] env[62952]: DEBUG nova.scheduler.client.report [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 926.693704] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263419, 'name': Rename_Task, 'duration_secs': 0.177294} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.693704] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 926.694360] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32f7b25a-f058-4353-bead-82c7c2bd4cff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.703172] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 926.703172] env[62952]: value = "task-1263422" [ 926.703172] env[62952]: _type = "Task" [ 926.703172] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.715544] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263422, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.809718] env[62952]: DEBUG nova.compute.manager [req-fe902502-07f4-420f-a9a3-dda61db3de4c req-d17b719a-e8fd-497a-873f-cb58b9ae9da3 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Received event network-changed-6bab534b-286c-4391-aeeb-fbb57655925a {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 926.810089] env[62952]: DEBUG nova.compute.manager [req-fe902502-07f4-420f-a9a3-dda61db3de4c req-d17b719a-e8fd-497a-873f-cb58b9ae9da3 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Refreshing instance network info cache due to event network-changed-6bab534b-286c-4391-aeeb-fbb57655925a. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 926.810488] env[62952]: DEBUG oslo_concurrency.lockutils [req-fe902502-07f4-420f-a9a3-dda61db3de4c req-d17b719a-e8fd-497a-873f-cb58b9ae9da3 service nova] Acquiring lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.823253] env[62952]: DEBUG nova.network.neutron [-] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.879271] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.879677] env[62952]: DEBUG nova.compute.manager [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Instance network_info: |[{"id": "6bab534b-286c-4391-aeeb-fbb57655925a", "address": "fa:16:3e:e2:f4:40", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bab534b-28", "ovs_interfaceid": "6bab534b-286c-4391-aeeb-fbb57655925a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 926.886226] env[62952]: DEBUG oslo_concurrency.lockutils [req-fe902502-07f4-420f-a9a3-dda61db3de4c req-d17b719a-e8fd-497a-873f-cb58b9ae9da3 service nova] Acquired lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.886226] env[62952]: DEBUG nova.network.neutron [req-fe902502-07f4-420f-a9a3-dda61db3de4c req-d17b719a-e8fd-497a-873f-cb58b9ae9da3 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Refreshing network info cache for port 6bab534b-286c-4391-aeeb-fbb57655925a {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 926.886226] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e2:f4:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '838c9497-35dd-415e-96c7-8dc21b0cd4b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6bab534b-286c-4391-aeeb-fbb57655925a', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 926.894427] env[62952]: DEBUG oslo.service.loopingcall [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.899201] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 926.900414] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': task-1263420, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.901130] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-de70dfbf-2b7e-4f33-848f-359cea895759 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.924116] env[62952]: DEBUG oslo_vmware.api [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263421, 'name': CreateSnapshot_Task, 'duration_secs': 0.517291} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.924116] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 926.924116] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58068e4-e728-4f6d-8d51-0fa10522b4d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.930461] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16255acf-23ae-4820-b27c-a9abac497d7b tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.930461] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16255acf-23ae-4820-b27c-a9abac497d7b tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.930461] env[62952]: DEBUG nova.compute.manager [None req-16255acf-23ae-4820-b27c-a9abac497d7b tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.930814] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 926.930814] env[62952]: value = "task-1263423" [ 926.930814] env[62952]: _type = "Task" [ 926.930814] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.936053] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0f5f630-41ff-4716-9f57-c691c7422f68 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.947267] env[62952]: DEBUG nova.compute.manager [None req-16255acf-23ae-4820-b27c-a9abac497d7b tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62952) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 926.948342] env[62952]: DEBUG nova.objects.instance [None req-16255acf-23ae-4820-b27c-a9abac497d7b tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lazy-loading 'flavor' on Instance uuid 6e289c20-8dd2-4680-b816-9bbe82fe7aec {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.953768] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263423, 'name': CreateVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.075864] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.046s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.079428] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.133s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.079700] env[62952]: DEBUG nova.objects.instance [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lazy-loading 'resources' on Instance uuid 58e703ea-7ccf-4114-a5a2-49533390ed37 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.110203] env[62952]: INFO nova.scheduler.client.report [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleted allocations for instance ba022bea-e357-4541-a573-d1da2eb49b46 [ 927.216269] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263422, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.326029] env[62952]: INFO nova.compute.manager [-] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Took 1.03 seconds to deallocate network for instance. [ 927.384329] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': task-1263420, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.456240] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 927.459037] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a2453a4b-2d36-40f4-a3e6-6e8debd9cb23 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.464485] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-16255acf-23ae-4820-b27c-a9abac497d7b tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 927.464485] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-82c85cde-ed2d-4705-a9a4-a88015cdd958 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.469433] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263423, 'name': CreateVM_Task, 'duration_secs': 0.388005} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.472268] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 927.472661] env[62952]: DEBUG oslo_vmware.api [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 927.472661] env[62952]: value = "task-1263424" [ 927.472661] env[62952]: _type = "Task" [ 927.472661] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.474382] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.474568] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.474953] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 927.475375] env[62952]: DEBUG oslo_vmware.api [None req-16255acf-23ae-4820-b27c-a9abac497d7b tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 927.475375] env[62952]: value = "task-1263425" [ 927.475375] env[62952]: _type = "Task" [ 927.475375] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.475660] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf123f13-6d4e-4bf8-b47b-ae7c1500cd62 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.493384] env[62952]: DEBUG oslo_vmware.api [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263424, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.498357] env[62952]: DEBUG oslo_vmware.api [None req-16255acf-23ae-4820-b27c-a9abac497d7b tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263425, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.498746] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 927.498746] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d5110c-dcfe-0813-189a-bc4a908a7bb8" [ 927.498746] env[62952]: _type = "Task" [ 927.498746] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.514902] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d5110c-dcfe-0813-189a-bc4a908a7bb8, 'name': SearchDatastore_Task, 'duration_secs': 0.011346} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.516347] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.516593] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 927.516845] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.516990] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.517181] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 927.517470] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b884367-1dc0-4464-90b9-0251d7bd17d3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.529831] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 927.530534] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 927.531398] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92cc2d4e-aeda-4c7d-8f5a-38d118d9215d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.540444] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 927.540444] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5278d8de-348c-62ef-d1b9-2774d8df4cbe" [ 927.540444] env[62952]: _type = "Task" [ 927.540444] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.555428] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5278d8de-348c-62ef-d1b9-2774d8df4cbe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.623823] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eaaf52b9-55fa-461c-b1c3-b8135dc831aa tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "ba022bea-e357-4541-a573-d1da2eb49b46" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.083s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.714386] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263422, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.841152] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.891537] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': task-1263420, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.075534} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.892572] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 927.892759] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93d3f439-5a5c-4a47-ae77-20ea995810ac {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.929247] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20/45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.932672] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-062dc32a-3ca6-46be-a5dd-303b9cd5c761 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.964110] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Waiting for the task: (returnval){ [ 927.964110] env[62952]: value = "task-1263426" [ 927.964110] env[62952]: _type = "Task" [ 927.964110] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.978388] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': task-1263426, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.006888] env[62952]: DEBUG oslo_vmware.api [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263424, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.008378] env[62952]: DEBUG oslo_vmware.api [None req-16255acf-23ae-4820-b27c-a9abac497d7b tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263425, 'name': PowerOffVM_Task, 'duration_secs': 0.399022} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.008378] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-16255acf-23ae-4820-b27c-a9abac497d7b tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 928.008378] env[62952]: DEBUG nova.compute.manager [None req-16255acf-23ae-4820-b27c-a9abac497d7b tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 928.009070] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c0dc3c-698a-47ab-bd3f-44d7f5179ffa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.040209] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd6a605-9874-420b-9539-6468f7d1dbd0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.055504] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5278d8de-348c-62ef-d1b9-2774d8df4cbe, 'name': SearchDatastore_Task, 'duration_secs': 0.047156} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.056660] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6201581-7bfe-4fb8-95e5-90ccf550c7a2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.063919] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f7975d0-cd68-4c21-ab6b-e9e41806a186 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.096725] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 928.096725] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525a15b5-1f6f-29d7-0412-4293f991e0fe" [ 928.096725] env[62952]: _type = "Task" [ 928.096725] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.100219] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2ef936-7fb3-4e3e-be19-b24986ffd34f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.121772] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525a15b5-1f6f-29d7-0412-4293f991e0fe, 'name': SearchDatastore_Task, 'duration_secs': 0.013757} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.121982] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.122269] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] dafa4926-3ab5-47f9-8ff2-4049e05c3113/dafa4926-3ab5-47f9-8ff2-4049e05c3113.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 928.127030] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec416c2d-7ba8-4f3c-ba63-9b8490ab5a72 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.132828] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-04f15e0d-2695-4219-9d9d-375f171da4e3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.147139] env[62952]: DEBUG nova.compute.provider_tree [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.151033] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 928.151033] env[62952]: value = "task-1263427" [ 928.151033] env[62952]: _type = "Task" [ 928.151033] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.160418] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263427, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.210901] env[62952]: DEBUG nova.network.neutron [req-fe902502-07f4-420f-a9a3-dda61db3de4c req-d17b719a-e8fd-497a-873f-cb58b9ae9da3 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Updated VIF entry in instance network info cache for port 6bab534b-286c-4391-aeeb-fbb57655925a. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 928.211383] env[62952]: DEBUG nova.network.neutron [req-fe902502-07f4-420f-a9a3-dda61db3de4c req-d17b719a-e8fd-497a-873f-cb58b9ae9da3 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Updating instance_info_cache with network_info: [{"id": "6bab534b-286c-4391-aeeb-fbb57655925a", "address": "fa:16:3e:e2:f4:40", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bab534b-28", "ovs_interfaceid": "6bab534b-286c-4391-aeeb-fbb57655925a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.218193] env[62952]: DEBUG oslo_vmware.api [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263422, 'name': PowerOnVM_Task, 'duration_secs': 1.286216} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.218543] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 928.218702] env[62952]: INFO nova.compute.manager [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Took 16.43 seconds to spawn the instance on the hypervisor. [ 928.218900] env[62952]: DEBUG nova.compute.manager [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 928.219787] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c2b75f6-7f28-4a97-8633-af7aa607bc2a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.480289] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': task-1263426, 'name': ReconfigVM_Task, 'duration_secs': 0.381226} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.480676] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20/45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 928.482303] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-99bd17cf-3e0b-487b-b6a8-9a1466e33761 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.494740] env[62952]: DEBUG oslo_vmware.api [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263424, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.496304] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Waiting for the task: (returnval){ [ 928.496304] env[62952]: value = "task-1263428" [ 928.496304] env[62952]: _type = "Task" [ 928.496304] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.506260] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': task-1263428, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.528395] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16255acf-23ae-4820-b27c-a9abac497d7b tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.598s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.651170] env[62952]: DEBUG nova.scheduler.client.report [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 928.674247] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263427, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507657} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.674546] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] dafa4926-3ab5-47f9-8ff2-4049e05c3113/dafa4926-3ab5-47f9-8ff2-4049e05c3113.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 928.676015] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 928.676015] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-214b6f7b-7061-47a2-bce3-96f1f2e5ca2a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.683850] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 928.683850] env[62952]: value = "task-1263429" [ 928.683850] env[62952]: _type = "Task" [ 928.683850] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.693824] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263429, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.715839] env[62952]: DEBUG oslo_concurrency.lockutils [req-fe902502-07f4-420f-a9a3-dda61db3de4c req-d17b719a-e8fd-497a-873f-cb58b9ae9da3 service nova] Releasing lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.738489] env[62952]: INFO nova.compute.manager [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Took 37.68 seconds to build instance. [ 928.997141] env[62952]: DEBUG oslo_vmware.api [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263424, 'name': CloneVM_Task, 'duration_secs': 1.353456} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.997141] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Created linked-clone VM from snapshot [ 928.997388] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a97958d-4ce3-4834-877d-eeceed216f2a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.009930] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Uploading image 37e6f880-0767-4129-8c81-837b59b5a494 {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 929.016147] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': task-1263428, 'name': Rename_Task, 'duration_secs': 0.156493} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.016147] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 929.016147] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-86cbac0e-a66a-40df-a253-31059e5c06f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.022676] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Waiting for the task: (returnval){ [ 929.022676] env[62952]: value = "task-1263430" [ 929.022676] env[62952]: _type = "Task" [ 929.022676] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.031156] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': task-1263430, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.040160] env[62952]: DEBUG oslo_vmware.rw_handles [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 929.040160] env[62952]: value = "vm-271908" [ 929.040160] env[62952]: _type = "VirtualMachine" [ 929.040160] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 929.040474] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ff3aa802-ab96-41bf-b212-472ca661f6f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.047073] env[62952]: DEBUG oslo_vmware.rw_handles [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lease: (returnval){ [ 929.047073] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52fee232-342b-b71c-7608-144d97711950" [ 929.047073] env[62952]: _type = "HttpNfcLease" [ 929.047073] env[62952]: } obtained for exporting VM: (result){ [ 929.047073] env[62952]: value = "vm-271908" [ 929.047073] env[62952]: _type = "VirtualMachine" [ 929.047073] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 929.047372] env[62952]: DEBUG oslo_vmware.api [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the lease: (returnval){ [ 929.047372] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52fee232-342b-b71c-7608-144d97711950" [ 929.047372] env[62952]: _type = "HttpNfcLease" [ 929.047372] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 929.053785] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 929.053785] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52fee232-342b-b71c-7608-144d97711950" [ 929.053785] env[62952]: _type = "HttpNfcLease" [ 929.053785] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 929.163069] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.084s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.165571] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.378s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.166994] env[62952]: DEBUG nova.objects.instance [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62952) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 929.194486] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263429, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.152696} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.194486] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 929.195183] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29db33f6-18d0-4a60-b4b0-3c6d92e24b19 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.221595] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] dafa4926-3ab5-47f9-8ff2-4049e05c3113/dafa4926-3ab5-47f9-8ff2-4049e05c3113.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 929.222699] env[62952]: INFO nova.scheduler.client.report [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Deleted allocations for instance 58e703ea-7ccf-4114-a5a2-49533390ed37 [ 929.223607] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3efda571-f4cb-4065-822e-211de364660e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.242729] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c5eccadf-8860-4f7d-977a-6ccefdf4e89d tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "7a73cb49-3e93-4fcc-b613-3808a8d5c64a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.390s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.251717] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 929.251717] env[62952]: value = "task-1263432" [ 929.251717] env[62952]: _type = "Task" [ 929.251717] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.262465] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263432, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.365197] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.365544] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.539473] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': task-1263430, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.555253] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 929.555253] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52fee232-342b-b71c-7608-144d97711950" [ 929.555253] env[62952]: _type = "HttpNfcLease" [ 929.555253] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 929.556311] env[62952]: DEBUG oslo_vmware.rw_handles [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 929.556311] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52fee232-342b-b71c-7608-144d97711950" [ 929.556311] env[62952]: _type = "HttpNfcLease" [ 929.556311] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 929.556685] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b4d8cf-2cd1-4899-a613-722f72dbc08e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.566063] env[62952]: DEBUG oslo_vmware.rw_handles [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dc52af-c1f1-b012-e41b-93e85db96928/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 929.566063] env[62952]: DEBUG oslo_vmware.rw_handles [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dc52af-c1f1-b012-e41b-93e85db96928/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 929.747489] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c333631c-54ad-4ece-aaec-8f2fe83b557f tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "58e703ea-7ccf-4114-a5a2-49533390ed37" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.808s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.759883] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-86ec31fa-bd5d-4f07-9967-6f6dd207fd86 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.766547] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263432, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.868031] env[62952]: DEBUG nova.compute.manager [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 930.031026] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "7a73cb49-3e93-4fcc-b613-3808a8d5c64a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.031026] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "7a73cb49-3e93-4fcc-b613-3808a8d5c64a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.031026] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "7a73cb49-3e93-4fcc-b613-3808a8d5c64a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.031200] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "7a73cb49-3e93-4fcc-b613-3808a8d5c64a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.031284] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "7a73cb49-3e93-4fcc-b613-3808a8d5c64a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.034816] env[62952]: INFO nova.compute.manager [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Terminating instance [ 930.043595] env[62952]: DEBUG nova.compute.manager [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 930.043595] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 930.044768] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3110d04-f0ea-4197-aa88-ee40828bf51f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.057060] env[62952]: DEBUG oslo_vmware.api [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': task-1263430, 'name': PowerOnVM_Task, 'duration_secs': 0.961928} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.057060] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 930.057259] env[62952]: INFO nova.compute.manager [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Took 9.62 seconds to spawn the instance on the hypervisor. [ 930.057575] env[62952]: DEBUG nova.compute.manager [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 930.057961] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 930.059137] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb530796-f857-49a4-ac20-837e0508227e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.063384] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c58f486-4bc9-452a-bcb2-82a26fd175e4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.076804] env[62952]: DEBUG oslo_vmware.api [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 930.076804] env[62952]: value = "task-1263433" [ 930.076804] env[62952]: _type = "Task" [ 930.076804] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.090963] env[62952]: DEBUG oslo_vmware.api [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263433, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.177763] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d01129-81ed-4238-bb1d-7bdfd211d3b3 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.179075] env[62952]: DEBUG oslo_concurrency.lockutils [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.538s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.179568] env[62952]: DEBUG nova.objects.instance [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lazy-loading 'resources' on Instance uuid 9cc7bc2f-ff22-43cf-be79-ff8c6426115b {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 930.263878] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263432, 'name': ReconfigVM_Task, 'duration_secs': 0.667208} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.268275] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Reconfigured VM instance instance-00000049 to attach disk [datastore1] dafa4926-3ab5-47f9-8ff2-4049e05c3113/dafa4926-3ab5-47f9-8ff2-4049e05c3113.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.269194] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6cf39819-d706-4248-abfa-de49958fc39b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.275195] env[62952]: DEBUG nova.objects.instance [None req-aa2aff5a-4863-41ea-9a23-5bf38503a431 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lazy-loading 'flavor' on Instance uuid 6e289c20-8dd2-4680-b816-9bbe82fe7aec {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 930.279060] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 930.279060] env[62952]: value = "task-1263434" [ 930.279060] env[62952]: _type = "Task" [ 930.279060] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.287926] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263434, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.400822] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.517171] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.518208] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.518208] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.518208] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.518208] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.526321] env[62952]: INFO nova.compute.manager [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Terminating instance [ 930.529064] env[62952]: DEBUG nova.compute.manager [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 930.529689] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 930.530962] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89658c01-9a5b-4842-bcf9-9e94a102c031 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.541034] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 930.542224] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7573e34a-8d5e-4274-8631-9999a8f85b33 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.550938] env[62952]: DEBUG oslo_vmware.api [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 930.550938] env[62952]: value = "task-1263435" [ 930.550938] env[62952]: _type = "Task" [ 930.550938] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.564612] env[62952]: DEBUG oslo_vmware.api [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263435, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.594339] env[62952]: INFO nova.compute.manager [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Took 33.03 seconds to build instance. [ 930.603891] env[62952]: DEBUG oslo_vmware.api [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263433, 'name': PowerOffVM_Task, 'duration_secs': 0.299942} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.604360] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 930.604711] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 930.609172] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-23af12ff-5080-4e9c-bc40-c89e1b2d553a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.672410] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 930.672636] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 930.672636] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Deleting the datastore file [datastore1] 7a73cb49-3e93-4fcc-b613-3808a8d5c64a {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 930.673309] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-facdcf51-cb7d-4a42-9a93-abeb0eddb02d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.685568] env[62952]: DEBUG nova.objects.instance [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lazy-loading 'numa_topology' on Instance uuid 9cc7bc2f-ff22-43cf-be79-ff8c6426115b {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 930.690036] env[62952]: DEBUG oslo_vmware.api [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 930.690036] env[62952]: value = "task-1263437" [ 930.690036] env[62952]: _type = "Task" [ 930.690036] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.711790] env[62952]: DEBUG oslo_vmware.api [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263437, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.788595] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa2aff5a-4863-41ea-9a23-5bf38503a431 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.788800] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa2aff5a-4863-41ea-9a23-5bf38503a431 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.788996] env[62952]: DEBUG nova.network.neutron [None req-aa2aff5a-4863-41ea-9a23-5bf38503a431 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 930.789254] env[62952]: DEBUG nova.objects.instance [None req-aa2aff5a-4863-41ea-9a23-5bf38503a431 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lazy-loading 'info_cache' on Instance uuid 6e289c20-8dd2-4680-b816-9bbe82fe7aec {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 930.797292] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263434, 'name': Rename_Task, 'duration_secs': 0.179886} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.801490] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 930.801809] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a91db178-dbcb-46ba-b6ee-79b87fcaebc7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.809305] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 930.809305] env[62952]: value = "task-1263438" [ 930.809305] env[62952]: _type = "Task" [ 930.809305] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.820985] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263438, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.063051] env[62952]: DEBUG oslo_vmware.api [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263435, 'name': PowerOffVM_Task, 'duration_secs': 0.271585} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.063340] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 931.063513] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 931.064172] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e60af519-083d-4d6e-8cf4-365c6ec188ae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.097667] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30526212-2e9b-41c6-9b95-9d0a49936f68 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Lock "45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.628s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.197207] env[62952]: DEBUG nova.objects.base [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Object Instance<9cc7bc2f-ff22-43cf-be79-ff8c6426115b> lazy-loaded attributes: resources,numa_topology {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 931.206536] env[62952]: DEBUG oslo_vmware.api [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263437, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.236856} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.206882] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 931.207403] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 931.207934] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 931.207934] env[62952]: INFO nova.compute.manager [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Took 1.16 seconds to destroy the instance on the hypervisor. [ 931.208302] env[62952]: DEBUG oslo.service.loopingcall [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.208888] env[62952]: DEBUG nova.compute.manager [-] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 931.208966] env[62952]: DEBUG nova.network.neutron [-] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 931.236263] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 931.236682] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 931.237046] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Deleting the datastore file [datastore2] c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 931.237678] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-98690e68-4035-4c4b-8429-e16b98dc04c7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.244992] env[62952]: DEBUG oslo_vmware.api [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for the task: (returnval){ [ 931.244992] env[62952]: value = "task-1263440" [ 931.244992] env[62952]: _type = "Task" [ 931.244992] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.260087] env[62952]: DEBUG oslo_vmware.api [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263440, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.292885] env[62952]: DEBUG nova.objects.base [None req-aa2aff5a-4863-41ea-9a23-5bf38503a431 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Object Instance<6e289c20-8dd2-4680-b816-9bbe82fe7aec> lazy-loaded attributes: flavor,info_cache {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 931.324503] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263438, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.587046] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c68d301-5e15-49b9-a29f-8960b5380bf1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.594951] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba1a2d0-1d30-4a63-a0c1-4153319d597b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.636435] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17556d19-002d-4b7a-8964-da97925397a7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.643347] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26af5d9f-8201-456c-9d08-19db7075d9fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.659525] env[62952]: DEBUG nova.compute.provider_tree [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.759600] env[62952]: DEBUG oslo_vmware.api [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Task: {'id': task-1263440, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.29369} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.760096] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 931.760269] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 931.760591] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 931.760850] env[62952]: INFO nova.compute.manager [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Took 1.23 seconds to destroy the instance on the hypervisor. [ 931.761166] env[62952]: DEBUG oslo.service.loopingcall [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.761428] env[62952]: DEBUG nova.compute.manager [-] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 931.761581] env[62952]: DEBUG nova.network.neutron [-] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 931.822161] env[62952]: DEBUG oslo_vmware.api [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263438, 'name': PowerOnVM_Task, 'duration_secs': 0.742121} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.822561] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 931.822842] env[62952]: INFO nova.compute.manager [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Took 8.80 seconds to spawn the instance on the hypervisor. [ 931.823059] env[62952]: DEBUG nova.compute.manager [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.823902] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7693514a-7dea-426e-a76e-6224a0c058d6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.897780] env[62952]: DEBUG oslo_concurrency.lockutils [None req-960016aa-e1cc-4da7-b4a2-f87e713e32a0 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Acquiring lock "interface-45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.897888] env[62952]: DEBUG oslo_concurrency.lockutils [None req-960016aa-e1cc-4da7-b4a2-f87e713e32a0 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Lock "interface-45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.898260] env[62952]: DEBUG nova.objects.instance [None req-960016aa-e1cc-4da7-b4a2-f87e713e32a0 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Lazy-loading 'flavor' on Instance uuid 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.169277] env[62952]: DEBUG nova.scheduler.client.report [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 932.175307] env[62952]: DEBUG nova.network.neutron [-] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.301210] env[62952]: DEBUG nova.network.neutron [None req-aa2aff5a-4863-41ea-9a23-5bf38503a431 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance_info_cache with network_info: [{"id": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "address": "fa:16:3e:eb:c2:8d", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e2e9d1e-dc", "ovs_interfaceid": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.345223] env[62952]: INFO nova.compute.manager [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Took 33.99 seconds to build instance. [ 932.409531] env[62952]: DEBUG nova.objects.instance [None req-960016aa-e1cc-4da7-b4a2-f87e713e32a0 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Lazy-loading 'pci_requests' on Instance uuid 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.553760] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "dbca4369-1ed3-493a-9847-9fa1e4293475" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.553986] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "dbca4369-1ed3-493a-9847-9fa1e4293475" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.613314] env[62952]: DEBUG nova.compute.manager [req-9c7143f4-5370-44bd-ba70-50bcc9f5bfe7 req-5ff0972e-ed82-490e-995d-abcf9f3c10bf service nova] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Received event network-vif-deleted-c6818fec-4bb8-4a9d-9106-6ef1f220347b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 932.680532] env[62952]: DEBUG oslo_concurrency.lockutils [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.501s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.683442] env[62952]: INFO nova.compute.manager [-] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Took 1.47 seconds to deallocate network for instance. [ 932.685391] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.242s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.685932] env[62952]: INFO nova.compute.claims [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 932.716803] env[62952]: DEBUG nova.network.neutron [-] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.804773] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa2aff5a-4863-41ea-9a23-5bf38503a431 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.847621] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b53235c-8188-4d40-91fd-fd6c262719d0 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.956s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.915046] env[62952]: DEBUG nova.objects.base [None req-960016aa-e1cc-4da7-b4a2-f87e713e32a0 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Object Instance<45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20> lazy-loaded attributes: flavor,pci_requests {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 932.915046] env[62952]: DEBUG nova.network.neutron [None req-960016aa-e1cc-4da7-b4a2-f87e713e32a0 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 933.057692] env[62952]: DEBUG nova.compute.manager [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 933.196585] env[62952]: DEBUG oslo_concurrency.lockutils [None req-099e7df6-fe34-4b34-805d-f2ecc9a0bb87 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 41.960s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.200270] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.200270] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 20.475s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.202262] env[62952]: INFO nova.compute.manager [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Unshelving [ 933.222021] env[62952]: INFO nova.compute.manager [-] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Took 1.46 seconds to deallocate network for instance. [ 933.263762] env[62952]: DEBUG oslo_concurrency.lockutils [None req-960016aa-e1cc-4da7-b4a2-f87e713e32a0 tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Lock "interface-45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.366s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.308432] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa2aff5a-4863-41ea-9a23-5bf38503a431 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 933.309349] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e0f0b29-464c-4c65-a033-ad6c6a2c28d4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.319564] env[62952]: DEBUG oslo_vmware.api [None req-aa2aff5a-4863-41ea-9a23-5bf38503a431 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 933.319564] env[62952]: value = "task-1263441" [ 933.319564] env[62952]: _type = "Task" [ 933.319564] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.329108] env[62952]: DEBUG oslo_vmware.api [None req-aa2aff5a-4863-41ea-9a23-5bf38503a431 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263441, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.582215] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.727710] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.831306] env[62952]: DEBUG oslo_vmware.api [None req-aa2aff5a-4863-41ea-9a23-5bf38503a431 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263441, 'name': PowerOnVM_Task, 'duration_secs': 0.424698} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.831518] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa2aff5a-4863-41ea-9a23-5bf38503a431 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 933.832165] env[62952]: DEBUG nova.compute.manager [None req-aa2aff5a-4863-41ea-9a23-5bf38503a431 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 933.832549] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3227cc-e8c4-44f5-9078-9e63c1bad3a0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.868232] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 933.868747] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 934.005984] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d151452c-d538-4b06-bccb-38f72db89ea6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.013744] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d222472-0f56-4d68-b503-d51e5a071054 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.045334] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d6ba39-da9d-439e-a36e-f243cae3a14a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.053500] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-437c576f-39b8-4bb2-82f2-d87b69aff4cd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.067410] env[62952]: DEBUG nova.compute.provider_tree [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.215341] env[62952]: DEBUG nova.compute.utils [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 934.377094] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 934.377443] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 934.377443] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 934.571033] env[62952]: DEBUG nova.scheduler.client.report [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 934.659709] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "77fe79de-8da8-486e-af63-91f9d8196ca9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.659933] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "77fe79de-8da8-486e-af63-91f9d8196ca9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.718975] env[62952]: INFO nova.virt.block_device [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Booting with volume 11fed6c5-a8bb-4173-ade8-22ca4aa0b85c at /dev/sdb [ 934.753367] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4e77d2d7-6740-41cb-84ef-0b964bf891c7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.763406] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920b6b0e-e44e-4da4-93eb-ef1663ad63ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.796305] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-809ee2f4-8793-499c-8d2b-473a177ecfd7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.805933] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5da32a2a-ebd4-4a02-9fed-44389c1d116f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.838400] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30602842-77bd-4c42-a68e-8457e635e175 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.845495] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954f2f96-4cdd-409f-af0b-afab865f6f1d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.858928] env[62952]: DEBUG nova.virt.block_device [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Updating existing volume attachment record: 302ecbb0-c1fd-465a-be1b-407a8b8594bb {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 934.884443] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 934.928994] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.931617] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquired lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.931617] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Forcefully refreshing network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 934.931617] env[62952]: DEBUG nova.objects.instance [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lazy-loading 'info_cache' on Instance uuid f4f8bee4-bf92-4321-afdc-16f7fb17dbae {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.082200] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.082200] env[62952]: DEBUG nova.compute.manager [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 935.085859] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.799s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.090035] env[62952]: DEBUG nova.objects.instance [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lazy-loading 'resources' on Instance uuid 3416eb38-b961-4d48-8b37-8b7b92375dc9 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.150029] env[62952]: DEBUG nova.compute.manager [req-7e8a5e30-cf8d-4d78-ad81-d22e1a8c4da6 req-b4fd2062-fce1-40c0-89d8-325ff8aa15e3 service nova] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Received event network-vif-deleted-52b68252-6346-4743-b795-f02ea8effef1 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 935.150240] env[62952]: DEBUG nova.compute.manager [req-7e8a5e30-cf8d-4d78-ad81-d22e1a8c4da6 req-b4fd2062-fce1-40c0-89d8-325ff8aa15e3 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Received event network-changed-b5b040c9-55a9-4929-b720-a1988f7492c9 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 935.151028] env[62952]: DEBUG nova.compute.manager [req-7e8a5e30-cf8d-4d78-ad81-d22e1a8c4da6 req-b4fd2062-fce1-40c0-89d8-325ff8aa15e3 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Refreshing instance network info cache due to event network-changed-b5b040c9-55a9-4929-b720-a1988f7492c9. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 935.151028] env[62952]: DEBUG oslo_concurrency.lockutils [req-7e8a5e30-cf8d-4d78-ad81-d22e1a8c4da6 req-b4fd2062-fce1-40c0-89d8-325ff8aa15e3 service nova] Acquiring lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.151028] env[62952]: DEBUG oslo_concurrency.lockutils [req-7e8a5e30-cf8d-4d78-ad81-d22e1a8c4da6 req-b4fd2062-fce1-40c0-89d8-325ff8aa15e3 service nova] Acquired lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.151028] env[62952]: DEBUG nova.network.neutron [req-7e8a5e30-cf8d-4d78-ad81-d22e1a8c4da6 req-b4fd2062-fce1-40c0-89d8-325ff8aa15e3 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Refreshing network info cache for port b5b040c9-55a9-4929-b720-a1988f7492c9 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 935.162483] env[62952]: DEBUG nova.compute.manager [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 935.232136] env[62952]: DEBUG oslo_concurrency.lockutils [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.232136] env[62952]: DEBUG oslo_concurrency.lockutils [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.400487] env[62952]: DEBUG oslo_concurrency.lockutils [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Acquiring lock "45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.400487] env[62952]: DEBUG oslo_concurrency.lockutils [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Lock "45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.400487] env[62952]: DEBUG oslo_concurrency.lockutils [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Acquiring lock "45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.400487] env[62952]: DEBUG oslo_concurrency.lockutils [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Lock "45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.400898] env[62952]: DEBUG oslo_concurrency.lockutils [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Lock "45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.402082] env[62952]: INFO nova.compute.manager [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Terminating instance [ 935.410262] env[62952]: DEBUG nova.compute.manager [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 935.410262] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 935.410262] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddded941-8708-4ee2-a80d-da2d87d06da1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.420756] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 935.420886] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e3d619d-0f1d-4ff1-b114-41757f3ff022 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.429707] env[62952]: DEBUG oslo_vmware.api [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Waiting for the task: (returnval){ [ 935.429707] env[62952]: value = "task-1263445" [ 935.429707] env[62952]: _type = "Task" [ 935.429707] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.444653] env[62952]: DEBUG oslo_vmware.api [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': task-1263445, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.590368] env[62952]: DEBUG nova.compute.utils [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 935.590368] env[62952]: DEBUG nova.compute.manager [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 935.590368] env[62952]: DEBUG nova.network.neutron [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 935.693493] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.724492] env[62952]: DEBUG nova.policy [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3ec1aecc0cf64aaea875a4280b114874', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fbb5ccb4d615429facd6932fb55b04e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 935.741195] env[62952]: INFO nova.compute.manager [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Detaching volume 42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d [ 935.811595] env[62952]: INFO nova.virt.block_device [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Attempting to driver detach volume 42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d from mountpoint /dev/sdb [ 935.811595] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 935.811892] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271870', 'volume_id': '42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d', 'name': 'volume-42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f4f8bee4-bf92-4321-afdc-16f7fb17dbae', 'attached_at': '', 'detached_at': '', 'volume_id': '42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d', 'serial': '42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 935.812725] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f95fb84e-07d3-40e7-bf39-33885adc67c0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.841654] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd58732-663a-49b0-98e8-68e10780fc7d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.852416] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a42200c6-2c81-422c-9842-c979dd3950b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.880206] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c58543f-81c4-46f8-b8b9-b06c083bf655 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.896893] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] The volume has not been displaced from its original location: [datastore1] volume-42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d/volume-42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 935.902620] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Reconfiguring VM instance instance-00000031 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 935.905967] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66b8755f-01c0-47e7-b409-8347ccc5d882 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.924970] env[62952]: DEBUG oslo_vmware.api [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 935.924970] env[62952]: value = "task-1263446" [ 935.924970] env[62952]: _type = "Task" [ 935.924970] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.943643] env[62952]: DEBUG oslo_vmware.api [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263446, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.947594] env[62952]: DEBUG oslo_vmware.api [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': task-1263445, 'name': PowerOffVM_Task, 'duration_secs': 0.278573} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.947852] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 935.948046] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 935.948762] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-545f29b4-d96a-4c87-8293-ae3bcb8c6d9f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.037079] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 936.037319] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 936.037745] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Deleting the datastore file [datastore2] 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 936.037831] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0cf2609b-e5d0-44a6-8a1a-67ed085e0259 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.047595] env[62952]: DEBUG oslo_vmware.api [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Waiting for the task: (returnval){ [ 936.047595] env[62952]: value = "task-1263448" [ 936.047595] env[62952]: _type = "Task" [ 936.047595] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.049057] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cbbc60b-2513-4cb7-981b-b9befa92ca88 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.063906] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b351bd9-f528-474e-ae9c-655ede6134ea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.067313] env[62952]: DEBUG oslo_vmware.api [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': task-1263448, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.097618] env[62952]: DEBUG nova.compute.manager [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 936.106127] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8641ba8f-1a01-4fda-bfb7-c785b732057a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.114924] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce181fc-663c-4532-a9c7-ba2f8c0393ac {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.129379] env[62952]: DEBUG nova.compute.provider_tree [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.320531] env[62952]: DEBUG nova.network.neutron [req-7e8a5e30-cf8d-4d78-ad81-d22e1a8c4da6 req-b4fd2062-fce1-40c0-89d8-325ff8aa15e3 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Updated VIF entry in instance network info cache for port b5b040c9-55a9-4929-b720-a1988f7492c9. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 936.321058] env[62952]: DEBUG nova.network.neutron [req-7e8a5e30-cf8d-4d78-ad81-d22e1a8c4da6 req-b4fd2062-fce1-40c0-89d8-325ff8aa15e3 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Updating instance_info_cache with network_info: [{"id": "b5b040c9-55a9-4929-b720-a1988f7492c9", "address": "fa:16:3e:0b:89:af", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5b040c9-55", "ovs_interfaceid": "b5b040c9-55a9-4929-b720-a1988f7492c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.350051] env[62952]: DEBUG nova.compute.manager [req-3502c916-789b-4a7e-a9d0-5403bb5c6f58 req-d116c7eb-d691-4aea-bbe2-cb58183a4868 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Received event network-changed-6bab534b-286c-4391-aeeb-fbb57655925a {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 936.350274] env[62952]: DEBUG nova.compute.manager [req-3502c916-789b-4a7e-a9d0-5403bb5c6f58 req-d116c7eb-d691-4aea-bbe2-cb58183a4868 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Refreshing instance network info cache due to event network-changed-6bab534b-286c-4391-aeeb-fbb57655925a. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 936.350510] env[62952]: DEBUG oslo_concurrency.lockutils [req-3502c916-789b-4a7e-a9d0-5403bb5c6f58 req-d116c7eb-d691-4aea-bbe2-cb58183a4868 service nova] Acquiring lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.350660] env[62952]: DEBUG oslo_concurrency.lockutils [req-3502c916-789b-4a7e-a9d0-5403bb5c6f58 req-d116c7eb-d691-4aea-bbe2-cb58183a4868 service nova] Acquired lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.350821] env[62952]: DEBUG nova.network.neutron [req-3502c916-789b-4a7e-a9d0-5403bb5c6f58 req-d116c7eb-d691-4aea-bbe2-cb58183a4868 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Refreshing network info cache for port 6bab534b-286c-4391-aeeb-fbb57655925a {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 936.435319] env[62952]: DEBUG oslo_vmware.api [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263446, 'name': ReconfigVM_Task, 'duration_secs': 0.442371} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.435673] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Reconfigured VM instance instance-00000031 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 936.440302] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e6229e74-4dd0-44c6-a868-bca3e6c785cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.455564] env[62952]: DEBUG oslo_vmware.api [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 936.455564] env[62952]: value = "task-1263449" [ 936.455564] env[62952]: _type = "Task" [ 936.455564] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.463700] env[62952]: DEBUG oslo_vmware.api [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263449, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.472862] env[62952]: DEBUG nova.network.neutron [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Successfully created port: d1e678fd-723e-413a-a8e8-44e88b909665 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 936.559408] env[62952]: DEBUG oslo_vmware.api [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Task: {'id': task-1263448, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.444392} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.559708] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 936.559899] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 936.560106] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 936.560314] env[62952]: INFO nova.compute.manager [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Took 1.15 seconds to destroy the instance on the hypervisor. [ 936.560610] env[62952]: DEBUG oslo.service.loopingcall [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 936.560822] env[62952]: DEBUG nova.compute.manager [-] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 936.560916] env[62952]: DEBUG nova.network.neutron [-] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 936.636221] env[62952]: DEBUG nova.scheduler.client.report [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 936.823909] env[62952]: DEBUG oslo_concurrency.lockutils [req-7e8a5e30-cf8d-4d78-ad81-d22e1a8c4da6 req-b4fd2062-fce1-40c0-89d8-325ff8aa15e3 service nova] Releasing lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.824296] env[62952]: DEBUG nova.compute.manager [req-7e8a5e30-cf8d-4d78-ad81-d22e1a8c4da6 req-b4fd2062-fce1-40c0-89d8-325ff8aa15e3 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Received event network-changed-6bab534b-286c-4391-aeeb-fbb57655925a {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 936.824455] env[62952]: DEBUG nova.compute.manager [req-7e8a5e30-cf8d-4d78-ad81-d22e1a8c4da6 req-b4fd2062-fce1-40c0-89d8-325ff8aa15e3 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Refreshing instance network info cache due to event network-changed-6bab534b-286c-4391-aeeb-fbb57655925a. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 936.824671] env[62952]: DEBUG oslo_concurrency.lockutils [req-7e8a5e30-cf8d-4d78-ad81-d22e1a8c4da6 req-b4fd2062-fce1-40c0-89d8-325ff8aa15e3 service nova] Acquiring lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.967454] env[62952]: DEBUG oslo_vmware.api [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263449, 'name': ReconfigVM_Task, 'duration_secs': 0.204507} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.967801] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271870', 'volume_id': '42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d', 'name': 'volume-42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f4f8bee4-bf92-4321-afdc-16f7fb17dbae', 'attached_at': '', 'detached_at': '', 'volume_id': '42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d', 'serial': '42ed3ad5-77bd-4f4f-8e0e-3a48f0c0f99d'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 937.055418] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Updating instance_info_cache with network_info: [{"id": "363ca83f-c6ac-4446-996c-edc459eb29bf", "address": "fa:16:3e:dd:39:7e", "network": {"id": "3c430a3e-4ea2-4c7b-8e05-733f4a306f04", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-441856160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8960909bb80f4395bb87b6de3c60ab2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap363ca83f-c6", "ovs_interfaceid": "363ca83f-c6ac-4446-996c-edc459eb29bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.070935] env[62952]: DEBUG nova.network.neutron [req-3502c916-789b-4a7e-a9d0-5403bb5c6f58 req-d116c7eb-d691-4aea-bbe2-cb58183a4868 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Updated VIF entry in instance network info cache for port 6bab534b-286c-4391-aeeb-fbb57655925a. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 937.070935] env[62952]: DEBUG nova.network.neutron [req-3502c916-789b-4a7e-a9d0-5403bb5c6f58 req-d116c7eb-d691-4aea-bbe2-cb58183a4868 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Updating instance_info_cache with network_info: [{"id": "6bab534b-286c-4391-aeeb-fbb57655925a", "address": "fa:16:3e:e2:f4:40", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bab534b-28", "ovs_interfaceid": "6bab534b-286c-4391-aeeb-fbb57655925a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.115241] env[62952]: DEBUG nova.compute.manager [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 937.146022] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.058s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.148381] env[62952]: DEBUG nova.virt.hardware [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 937.148469] env[62952]: DEBUG nova.virt.hardware [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 937.148857] env[62952]: DEBUG nova.virt.hardware [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 937.149081] env[62952]: DEBUG nova.virt.hardware [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 937.149845] env[62952]: DEBUG nova.virt.hardware [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 937.150095] env[62952]: DEBUG nova.virt.hardware [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 937.150919] env[62952]: DEBUG nova.virt.hardware [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 937.151131] env[62952]: DEBUG nova.virt.hardware [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 937.151315] env[62952]: DEBUG nova.virt.hardware [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 937.151479] env[62952]: DEBUG nova.virt.hardware [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 937.151653] env[62952]: DEBUG nova.virt.hardware [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 937.152256] env[62952]: DEBUG oslo_concurrency.lockutils [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.149s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.154073] env[62952]: INFO nova.compute.claims [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 937.162037] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aadbcbf9-31bc-415f-bccd-88233fcfa6bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.169044] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-016b41c4-f041-437b-a0d6-b0baafc93cdf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.185382] env[62952]: INFO nova.scheduler.client.report [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleted allocations for instance 3416eb38-b961-4d48-8b37-8b7b92375dc9 [ 937.308794] env[62952]: DEBUG nova.network.neutron [-] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.513893] env[62952]: DEBUG nova.objects.instance [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lazy-loading 'flavor' on Instance uuid f4f8bee4-bf92-4321-afdc-16f7fb17dbae {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.560620] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Releasing lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.560620] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Updated the network info_cache for instance {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 937.560620] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 937.560620] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 937.560620] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 937.560620] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 937.560851] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 937.560851] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 937.560851] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 937.560851] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 937.576250] env[62952]: DEBUG oslo_concurrency.lockutils [req-3502c916-789b-4a7e-a9d0-5403bb5c6f58 req-d116c7eb-d691-4aea-bbe2-cb58183a4868 service nova] Releasing lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.576250] env[62952]: DEBUG oslo_concurrency.lockutils [req-7e8a5e30-cf8d-4d78-ad81-d22e1a8c4da6 req-b4fd2062-fce1-40c0-89d8-325ff8aa15e3 service nova] Acquired lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.576250] env[62952]: DEBUG nova.network.neutron [req-7e8a5e30-cf8d-4d78-ad81-d22e1a8c4da6 req-b4fd2062-fce1-40c0-89d8-325ff8aa15e3 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Refreshing network info cache for port 6bab534b-286c-4391-aeeb-fbb57655925a {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 937.692459] env[62952]: DEBUG oslo_concurrency.lockutils [None req-58ccde16-a9e0-4df3-b2f3-213a2165d27e tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "3416eb38-b961-4d48-8b37-8b7b92375dc9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.992s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.811609] env[62952]: INFO nova.compute.manager [-] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Took 1.25 seconds to deallocate network for instance. [ 938.063136] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.294208] env[62952]: DEBUG nova.network.neutron [req-7e8a5e30-cf8d-4d78-ad81-d22e1a8c4da6 req-b4fd2062-fce1-40c0-89d8-325ff8aa15e3 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Updated VIF entry in instance network info cache for port 6bab534b-286c-4391-aeeb-fbb57655925a. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 938.294768] env[62952]: DEBUG nova.network.neutron [req-7e8a5e30-cf8d-4d78-ad81-d22e1a8c4da6 req-b4fd2062-fce1-40c0-89d8-325ff8aa15e3 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Updating instance_info_cache with network_info: [{"id": "6bab534b-286c-4391-aeeb-fbb57655925a", "address": "fa:16:3e:e2:f4:40", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bab534b-28", "ovs_interfaceid": "6bab534b-286c-4391-aeeb-fbb57655925a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.318558] env[62952]: DEBUG oslo_concurrency.lockutils [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.470014] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2b679b0f-d697-47fc-8dc8-ad095bd1c684 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.475962] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04fb613a-f958-4168-8624-68b94bb454ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.486025] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1e9e54-947d-42e5-a82d-bcb109577332 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.519265] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c860184-4af9-41c6-b0dc-7c2870d70673 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.525941] env[62952]: DEBUG oslo_concurrency.lockutils [None req-47287186-8fe8-4e41-8309-93800deccb12 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.295s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.527143] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2b679b0f-d697-47fc-8dc8-ad095bd1c684 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.058s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.527404] env[62952]: DEBUG nova.compute.manager [None req-2b679b0f-d697-47fc-8dc8-ad095bd1c684 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 938.528653] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-710bdcc4-b441-4286-b048-b0c9ad65d0cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.533317] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e2322a0-f235-4bd7-b5fa-606af2d89830 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.548327] env[62952]: DEBUG nova.compute.provider_tree [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.550067] env[62952]: DEBUG nova.compute.manager [None req-2b679b0f-d697-47fc-8dc8-ad095bd1c684 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62952) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 938.550693] env[62952]: DEBUG nova.objects.instance [None req-2b679b0f-d697-47fc-8dc8-ad095bd1c684 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lazy-loading 'flavor' on Instance uuid f4f8bee4-bf92-4321-afdc-16f7fb17dbae {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.712642] env[62952]: DEBUG nova.compute.manager [req-ec4dfee9-8a5e-47f0-bdcd-59a7b1c08d31 req-ba182c50-4c6d-4a68-9a2d-e14ceaff79c4 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Received event network-changed-b5b040c9-55a9-4929-b720-a1988f7492c9 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 938.712642] env[62952]: DEBUG nova.compute.manager [req-ec4dfee9-8a5e-47f0-bdcd-59a7b1c08d31 req-ba182c50-4c6d-4a68-9a2d-e14ceaff79c4 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Refreshing instance network info cache due to event network-changed-b5b040c9-55a9-4929-b720-a1988f7492c9. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 938.712642] env[62952]: DEBUG oslo_concurrency.lockutils [req-ec4dfee9-8a5e-47f0-bdcd-59a7b1c08d31 req-ba182c50-4c6d-4a68-9a2d-e14ceaff79c4 service nova] Acquiring lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.712750] env[62952]: DEBUG oslo_concurrency.lockutils [req-ec4dfee9-8a5e-47f0-bdcd-59a7b1c08d31 req-ba182c50-4c6d-4a68-9a2d-e14ceaff79c4 service nova] Acquired lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.712856] env[62952]: DEBUG nova.network.neutron [req-ec4dfee9-8a5e-47f0-bdcd-59a7b1c08d31 req-ba182c50-4c6d-4a68-9a2d-e14ceaff79c4 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Refreshing network info cache for port b5b040c9-55a9-4929-b720-a1988f7492c9 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 938.790797] env[62952]: DEBUG nova.compute.manager [req-8b5e34a0-c230-412c-bf52-731e3b7b42c5 req-ad9f121e-a4ba-4aa4-b37f-6c7819e6b682 service nova] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Received event network-vif-plugged-d1e678fd-723e-413a-a8e8-44e88b909665 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 938.791115] env[62952]: DEBUG oslo_concurrency.lockutils [req-8b5e34a0-c230-412c-bf52-731e3b7b42c5 req-ad9f121e-a4ba-4aa4-b37f-6c7819e6b682 service nova] Acquiring lock "e17b3d83-7cea-4a65-9457-7e57a6e7641a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.791334] env[62952]: DEBUG oslo_concurrency.lockutils [req-8b5e34a0-c230-412c-bf52-731e3b7b42c5 req-ad9f121e-a4ba-4aa4-b37f-6c7819e6b682 service nova] Lock "e17b3d83-7cea-4a65-9457-7e57a6e7641a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.791501] env[62952]: DEBUG oslo_concurrency.lockutils [req-8b5e34a0-c230-412c-bf52-731e3b7b42c5 req-ad9f121e-a4ba-4aa4-b37f-6c7819e6b682 service nova] Lock "e17b3d83-7cea-4a65-9457-7e57a6e7641a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.791692] env[62952]: DEBUG nova.compute.manager [req-8b5e34a0-c230-412c-bf52-731e3b7b42c5 req-ad9f121e-a4ba-4aa4-b37f-6c7819e6b682 service nova] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] No waiting events found dispatching network-vif-plugged-d1e678fd-723e-413a-a8e8-44e88b909665 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 938.791828] env[62952]: WARNING nova.compute.manager [req-8b5e34a0-c230-412c-bf52-731e3b7b42c5 req-ad9f121e-a4ba-4aa4-b37f-6c7819e6b682 service nova] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Received unexpected event network-vif-plugged-d1e678fd-723e-413a-a8e8-44e88b909665 for instance with vm_state building and task_state spawning. [ 938.796218] env[62952]: DEBUG nova.network.neutron [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Successfully updated port: d1e678fd-723e-413a-a8e8-44e88b909665 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 938.800124] env[62952]: DEBUG oslo_concurrency.lockutils [req-7e8a5e30-cf8d-4d78-ad81-d22e1a8c4da6 req-b4fd2062-fce1-40c0-89d8-325ff8aa15e3 service nova] Releasing lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.054635] env[62952]: DEBUG nova.scheduler.client.report [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.062758] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b679b0f-d697-47fc-8dc8-ad095bd1c684 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 939.062758] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8c257f7-c9de-4892-97ee-b469e0ce1fdf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.070079] env[62952]: DEBUG oslo_vmware.api [None req-2b679b0f-d697-47fc-8dc8-ad095bd1c684 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 939.070079] env[62952]: value = "task-1263451" [ 939.070079] env[62952]: _type = "Task" [ 939.070079] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.079991] env[62952]: DEBUG oslo_vmware.api [None req-2b679b0f-d697-47fc-8dc8-ad095bd1c684 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263451, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.301611] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "refresh_cache-e17b3d83-7cea-4a65-9457-7e57a6e7641a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.301765] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "refresh_cache-e17b3d83-7cea-4a65-9457-7e57a6e7641a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.301916] env[62952]: DEBUG nova.network.neutron [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 939.441781] env[62952]: DEBUG nova.network.neutron [req-ec4dfee9-8a5e-47f0-bdcd-59a7b1c08d31 req-ba182c50-4c6d-4a68-9a2d-e14ceaff79c4 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Updated VIF entry in instance network info cache for port b5b040c9-55a9-4929-b720-a1988f7492c9. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 939.442173] env[62952]: DEBUG nova.network.neutron [req-ec4dfee9-8a5e-47f0-bdcd-59a7b1c08d31 req-ba182c50-4c6d-4a68-9a2d-e14ceaff79c4 service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Updating instance_info_cache with network_info: [{"id": "b5b040c9-55a9-4929-b720-a1988f7492c9", "address": "fa:16:3e:0b:89:af", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5b040c9-55", "ovs_interfaceid": "b5b040c9-55a9-4929-b720-a1988f7492c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.559924] env[62952]: DEBUG oslo_concurrency.lockutils [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.560500] env[62952]: DEBUG nova.compute.manager [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 939.563729] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.007s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.563921] env[62952]: DEBUG nova.objects.instance [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62952) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 939.579362] env[62952]: DEBUG oslo_vmware.api [None req-2b679b0f-d697-47fc-8dc8-ad095bd1c684 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263451, 'name': PowerOffVM_Task, 'duration_secs': 0.410478} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.579612] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b679b0f-d697-47fc-8dc8-ad095bd1c684 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 939.579798] env[62952]: DEBUG nova.compute.manager [None req-2b679b0f-d697-47fc-8dc8-ad095bd1c684 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 939.580622] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33446b95-3f71-4fca-93a1-48bea7c4314d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.748057] env[62952]: DEBUG oslo_vmware.rw_handles [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dc52af-c1f1-b012-e41b-93e85db96928/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 939.749382] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b24f3dd-0fb1-4891-a2bd-59948c3309fb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.758200] env[62952]: DEBUG oslo_vmware.rw_handles [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dc52af-c1f1-b012-e41b-93e85db96928/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 939.758200] env[62952]: ERROR oslo_vmware.rw_handles [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dc52af-c1f1-b012-e41b-93e85db96928/disk-0.vmdk due to incomplete transfer. [ 939.758200] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-22d45251-f831-4a27-ba83-8b0bbb615b2f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.767285] env[62952]: DEBUG oslo_vmware.rw_handles [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dc52af-c1f1-b012-e41b-93e85db96928/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 939.767285] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Uploaded image 37e6f880-0767-4129-8c81-837b59b5a494 to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 939.768607] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 939.768967] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-be15e725-eb0c-4eab-bed1-adf210526b18 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.778027] env[62952]: DEBUG oslo_vmware.api [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 939.778027] env[62952]: value = "task-1263452" [ 939.778027] env[62952]: _type = "Task" [ 939.778027] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.787658] env[62952]: DEBUG oslo_vmware.api [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263452, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.839300] env[62952]: DEBUG nova.network.neutron [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 939.944598] env[62952]: DEBUG oslo_concurrency.lockutils [req-ec4dfee9-8a5e-47f0-bdcd-59a7b1c08d31 req-ba182c50-4c6d-4a68-9a2d-e14ceaff79c4 service nova] Releasing lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.945015] env[62952]: DEBUG nova.compute.manager [req-ec4dfee9-8a5e-47f0-bdcd-59a7b1c08d31 req-ba182c50-4c6d-4a68-9a2d-e14ceaff79c4 service nova] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Received event network-vif-deleted-a2903c7d-4cae-40ce-8f1a-52d2ea6a9bfd {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 940.068502] env[62952]: DEBUG nova.compute.utils [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 940.076338] env[62952]: DEBUG nova.compute.manager [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 940.076338] env[62952]: DEBUG nova.network.neutron [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 940.095982] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2b679b0f-d697-47fc-8dc8-ad095bd1c684 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.569s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.110739] env[62952]: DEBUG nova.network.neutron [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Updating instance_info_cache with network_info: [{"id": "d1e678fd-723e-413a-a8e8-44e88b909665", "address": "fa:16:3e:b4:17:22", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1e678fd-72", "ovs_interfaceid": "d1e678fd-723e-413a-a8e8-44e88b909665", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.147524] env[62952]: DEBUG nova.policy [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7eadc05f193645e6b6e86e9cdbbbaa71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '076fb3c53420447caf047f8422435152', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 940.287237] env[62952]: DEBUG oslo_vmware.api [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263452, 'name': Destroy_Task} progress is 33%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.488497] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.547149] env[62952]: DEBUG nova.network.neutron [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Successfully created port: 8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 940.576549] env[62952]: DEBUG nova.compute.manager [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 940.580186] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d8caaad-5550-4a66-b9fb-e61e1c89c325 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.581378] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.107s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.581617] env[62952]: DEBUG nova.objects.instance [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lazy-loading 'resources' on Instance uuid 051f9279-533a-4545-b9e2-272bf40956f6 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.613934] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "refresh_cache-e17b3d83-7cea-4a65-9457-7e57a6e7641a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.614413] env[62952]: DEBUG nova.compute.manager [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Instance network_info: |[{"id": "d1e678fd-723e-413a-a8e8-44e88b909665", "address": "fa:16:3e:b4:17:22", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1e678fd-72", "ovs_interfaceid": "d1e678fd-723e-413a-a8e8-44e88b909665", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 940.615625] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:17:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5ad8894f-e240-4013-8272-4e79daea0751', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd1e678fd-723e-413a-a8e8-44e88b909665', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 940.626575] env[62952]: DEBUG oslo.service.loopingcall [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.627691] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 940.628032] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc7e59c0-e7c2-4a69-abcc-00fcc4fd6cb4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.649825] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 940.649825] env[62952]: value = "task-1263453" [ 940.649825] env[62952]: _type = "Task" [ 940.649825] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.661149] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263453, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.789562] env[62952]: DEBUG oslo_vmware.api [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263452, 'name': Destroy_Task, 'duration_secs': 0.943376} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.790073] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Destroyed the VM [ 940.790073] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 940.790223] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d4c3a352-7972-4fd3-871f-c81edae2af59 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.796453] env[62952]: DEBUG oslo_vmware.api [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 940.796453] env[62952]: value = "task-1263454" [ 940.796453] env[62952]: _type = "Task" [ 940.796453] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.815149] env[62952]: DEBUG oslo_vmware.api [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263454, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.829765] env[62952]: DEBUG nova.compute.manager [req-1fb56e85-e48d-430f-8ef1-e5ee9d7eadfe req-bcb949b2-d050-4062-bfa7-557f37a1ed93 service nova] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Received event network-changed-d1e678fd-723e-413a-a8e8-44e88b909665 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 940.829975] env[62952]: DEBUG nova.compute.manager [req-1fb56e85-e48d-430f-8ef1-e5ee9d7eadfe req-bcb949b2-d050-4062-bfa7-557f37a1ed93 service nova] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Refreshing instance network info cache due to event network-changed-d1e678fd-723e-413a-a8e8-44e88b909665. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 940.830443] env[62952]: DEBUG oslo_concurrency.lockutils [req-1fb56e85-e48d-430f-8ef1-e5ee9d7eadfe req-bcb949b2-d050-4062-bfa7-557f37a1ed93 service nova] Acquiring lock "refresh_cache-e17b3d83-7cea-4a65-9457-7e57a6e7641a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.830607] env[62952]: DEBUG oslo_concurrency.lockutils [req-1fb56e85-e48d-430f-8ef1-e5ee9d7eadfe req-bcb949b2-d050-4062-bfa7-557f37a1ed93 service nova] Acquired lock "refresh_cache-e17b3d83-7cea-4a65-9457-7e57a6e7641a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.830829] env[62952]: DEBUG nova.network.neutron [req-1fb56e85-e48d-430f-8ef1-e5ee9d7eadfe req-bcb949b2-d050-4062-bfa7-557f37a1ed93 service nova] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Refreshing network info cache for port d1e678fd-723e-413a-a8e8-44e88b909665 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 941.164695] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263453, 'name': CreateVM_Task, 'duration_secs': 0.38009} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.164886] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 941.165719] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.165912] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.166267] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 941.167054] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74c7b824-8dde-480b-8452-538648d5bea9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.173083] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 941.173083] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520f916d-a306-6019-2658-905928cac0c9" [ 941.173083] env[62952]: _type = "Task" [ 941.173083] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.182891] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520f916d-a306-6019-2658-905928cac0c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.305966] env[62952]: DEBUG oslo_vmware.api [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263454, 'name': RemoveSnapshot_Task} progress is 56%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.316538] env[62952]: DEBUG nova.objects.instance [None req-91be5aa5-a3bc-49c4-8a22-cc99894849ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lazy-loading 'flavor' on Instance uuid f4f8bee4-bf92-4321-afdc-16f7fb17dbae {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.389389] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd31a93-3872-4859-ac42-a8f3bf9dbe9a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.398111] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9d349d-3381-46f3-828b-8bfe7c91c4e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.431901] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba10ccb8-8ad5-41ee-914d-dd7cea427bcc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.439361] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09aa439e-fa67-410e-ada2-88fe2c615206 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.452815] env[62952]: DEBUG nova.compute.provider_tree [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.573521] env[62952]: DEBUG nova.network.neutron [req-1fb56e85-e48d-430f-8ef1-e5ee9d7eadfe req-bcb949b2-d050-4062-bfa7-557f37a1ed93 service nova] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Updated VIF entry in instance network info cache for port d1e678fd-723e-413a-a8e8-44e88b909665. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 941.573941] env[62952]: DEBUG nova.network.neutron [req-1fb56e85-e48d-430f-8ef1-e5ee9d7eadfe req-bcb949b2-d050-4062-bfa7-557f37a1ed93 service nova] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Updating instance_info_cache with network_info: [{"id": "d1e678fd-723e-413a-a8e8-44e88b909665", "address": "fa:16:3e:b4:17:22", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1e678fd-72", "ovs_interfaceid": "d1e678fd-723e-413a-a8e8-44e88b909665", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.588049] env[62952]: DEBUG nova.compute.manager [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 941.613052] env[62952]: DEBUG nova.virt.hardware [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 941.613358] env[62952]: DEBUG nova.virt.hardware [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 941.613526] env[62952]: DEBUG nova.virt.hardware [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 941.613711] env[62952]: DEBUG nova.virt.hardware [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 941.613865] env[62952]: DEBUG nova.virt.hardware [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 941.614024] env[62952]: DEBUG nova.virt.hardware [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 941.614243] env[62952]: DEBUG nova.virt.hardware [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 941.614406] env[62952]: DEBUG nova.virt.hardware [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 941.614578] env[62952]: DEBUG nova.virt.hardware [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 941.614743] env[62952]: DEBUG nova.virt.hardware [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 941.614919] env[62952]: DEBUG nova.virt.hardware [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 941.615821] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23e89cb9-7fd6-41cd-bd92-9be75e3eaae4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.623453] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6897ae4-f9ec-451e-9070-a4af3e69d053 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.685226] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520f916d-a306-6019-2658-905928cac0c9, 'name': SearchDatastore_Task, 'duration_secs': 0.020131} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.685747] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.685927] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.686139] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.686313] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.686503] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 941.686783] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a20acfb-c9d4-40e0-833a-ed795cfcacfe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.695833] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 941.696016] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 941.696777] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-235fd3d9-4eb2-4127-a76a-ae31f69c3521 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.701686] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 941.701686] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520e6916-dedf-8c1c-2273-7a4806f5e82d" [ 941.701686] env[62952]: _type = "Task" [ 941.701686] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.709295] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520e6916-dedf-8c1c-2273-7a4806f5e82d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.810125] env[62952]: DEBUG oslo_vmware.api [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263454, 'name': RemoveSnapshot_Task, 'duration_secs': 0.669555} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.810125] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 941.810125] env[62952]: INFO nova.compute.manager [None req-1eb5a128-f30e-479c-b249-c7932c2105d7 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Took 16.00 seconds to snapshot the instance on the hypervisor. [ 941.825955] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91be5aa5-a3bc-49c4-8a22-cc99894849ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.825955] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91be5aa5-a3bc-49c4-8a22-cc99894849ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquired lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.825955] env[62952]: DEBUG nova.network.neutron [None req-91be5aa5-a3bc-49c4-8a22-cc99894849ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 941.825955] env[62952]: DEBUG nova.objects.instance [None req-91be5aa5-a3bc-49c4-8a22-cc99894849ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lazy-loading 'info_cache' on Instance uuid f4f8bee4-bf92-4321-afdc-16f7fb17dbae {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.957946] env[62952]: DEBUG nova.scheduler.client.report [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 942.076510] env[62952]: DEBUG oslo_concurrency.lockutils [req-1fb56e85-e48d-430f-8ef1-e5ee9d7eadfe req-bcb949b2-d050-4062-bfa7-557f37a1ed93 service nova] Releasing lock "refresh_cache-e17b3d83-7cea-4a65-9457-7e57a6e7641a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.172564] env[62952]: DEBUG nova.network.neutron [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Successfully updated port: 8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 942.213269] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520e6916-dedf-8c1c-2273-7a4806f5e82d, 'name': SearchDatastore_Task, 'duration_secs': 0.018511} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.214329] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87e73a00-2ef1-4a80-9d49-86f8e1a39ea8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.219603] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 942.219603] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5212946e-b920-24f0-b864-d29bbc870d76" [ 942.219603] env[62952]: _type = "Task" [ 942.219603] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.227046] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5212946e-b920-24f0-b864-d29bbc870d76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.328860] env[62952]: DEBUG nova.objects.base [None req-91be5aa5-a3bc-49c4-8a22-cc99894849ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 942.364540] env[62952]: DEBUG nova.compute.manager [req-fe437cc0-5de0-4733-9484-9ea3a577d22f req-12cf5eef-8e5d-4de6-8bf1-895c1fea2e0a service nova] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Received event network-vif-plugged-8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 942.364768] env[62952]: DEBUG oslo_concurrency.lockutils [req-fe437cc0-5de0-4733-9484-9ea3a577d22f req-12cf5eef-8e5d-4de6-8bf1-895c1fea2e0a service nova] Acquiring lock "20ea323b-8961-4981-8045-f9cb85815d99-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.365060] env[62952]: DEBUG oslo_concurrency.lockutils [req-fe437cc0-5de0-4733-9484-9ea3a577d22f req-12cf5eef-8e5d-4de6-8bf1-895c1fea2e0a service nova] Lock "20ea323b-8961-4981-8045-f9cb85815d99-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.365260] env[62952]: DEBUG oslo_concurrency.lockutils [req-fe437cc0-5de0-4733-9484-9ea3a577d22f req-12cf5eef-8e5d-4de6-8bf1-895c1fea2e0a service nova] Lock "20ea323b-8961-4981-8045-f9cb85815d99-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.365453] env[62952]: DEBUG nova.compute.manager [req-fe437cc0-5de0-4733-9484-9ea3a577d22f req-12cf5eef-8e5d-4de6-8bf1-895c1fea2e0a service nova] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] No waiting events found dispatching network-vif-plugged-8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 942.365581] env[62952]: WARNING nova.compute.manager [req-fe437cc0-5de0-4733-9484-9ea3a577d22f req-12cf5eef-8e5d-4de6-8bf1-895c1fea2e0a service nova] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Received unexpected event network-vif-plugged-8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855 for instance with vm_state building and task_state spawning. [ 942.464147] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.883s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.466513] env[62952]: DEBUG oslo_concurrency.lockutils [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.665s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.466747] env[62952]: DEBUG nova.objects.instance [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lazy-loading 'resources' on Instance uuid e5e7b2b5-34cf-4d37-87ff-1b57264a3516 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.473021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "interface-0533ee81-b69d-4f4e-9832-78c5ab4512b1-bdfdd199-b7c7-47fd-bcd7-15760750128b" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.473272] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "interface-0533ee81-b69d-4f4e-9832-78c5ab4512b1-bdfdd199-b7c7-47fd-bcd7-15760750128b" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.473616] env[62952]: DEBUG nova.objects.instance [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lazy-loading 'flavor' on Instance uuid 0533ee81-b69d-4f4e-9832-78c5ab4512b1 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.497238] env[62952]: INFO nova.scheduler.client.report [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Deleted allocations for instance 051f9279-533a-4545-b9e2-272bf40956f6 [ 942.671942] env[62952]: DEBUG oslo_concurrency.lockutils [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "refresh_cache-20ea323b-8961-4981-8045-f9cb85815d99" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.672252] env[62952]: DEBUG oslo_concurrency.lockutils [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "refresh_cache-20ea323b-8961-4981-8045-f9cb85815d99" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.672287] env[62952]: DEBUG nova.network.neutron [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 942.729959] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5212946e-b920-24f0-b864-d29bbc870d76, 'name': SearchDatastore_Task, 'duration_secs': 0.080072} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.730247] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.730503] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] e17b3d83-7cea-4a65-9457-7e57a6e7641a/e17b3d83-7cea-4a65-9457-7e57a6e7641a.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 942.730783] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-51f18222-71b6-44c4-b2f3-0a5119658b59 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.737142] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 942.737142] env[62952]: value = "task-1263455" [ 942.737142] env[62952]: _type = "Task" [ 942.737142] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.744631] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263455, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.008846] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a8b47c82-c3f2-4b9f-b54f-8814debfb132 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "051f9279-533a-4545-b9e2-272bf40956f6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.531s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.126540] env[62952]: DEBUG nova.network.neutron [None req-91be5aa5-a3bc-49c4-8a22-cc99894849ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Updating instance_info_cache with network_info: [{"id": "363ca83f-c6ac-4446-996c-edc459eb29bf", "address": "fa:16:3e:dd:39:7e", "network": {"id": "3c430a3e-4ea2-4c7b-8e05-733f4a306f04", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-441856160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8960909bb80f4395bb87b6de3c60ab2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap363ca83f-c6", "ovs_interfaceid": "363ca83f-c6ac-4446-996c-edc459eb29bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.177792] env[62952]: DEBUG nova.objects.instance [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lazy-loading 'pci_requests' on Instance uuid 0533ee81-b69d-4f4e-9832-78c5ab4512b1 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.210307] env[62952]: DEBUG nova.network.neutron [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 943.249824] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263455, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.443122} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.252022] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] e17b3d83-7cea-4a65-9457-7e57a6e7641a/e17b3d83-7cea-4a65-9457-7e57a6e7641a.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 943.254684] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 943.255983] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-be86d80a-43b9-4ef7-9cb4-17575ba438e7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.269021] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 943.269021] env[62952]: value = "task-1263456" [ 943.269021] env[62952]: _type = "Task" [ 943.269021] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.278824] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263456, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.343108] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aabb85ed-2411-4e32-8dd2-e458adc3888a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.350727] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0838407c-5fc5-4417-b9bb-d7020556f592 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.387673] env[62952]: DEBUG nova.network.neutron [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Updating instance_info_cache with network_info: [{"id": "8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855", "address": "fa:16:3e:65:72:4d", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ce9dcc3-e5", "ovs_interfaceid": "8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.390890] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e908b96-52e4-4720-8107-12ec78208ac2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.399145] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca85d05-0a2c-477a-99a2-0cee4ece9589 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.414399] env[62952]: DEBUG nova.compute.provider_tree [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.629752] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91be5aa5-a3bc-49c4-8a22-cc99894849ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Releasing lock "refresh_cache-f4f8bee4-bf92-4321-afdc-16f7fb17dbae" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.681620] env[62952]: DEBUG nova.objects.base [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Object Instance<0533ee81-b69d-4f4e-9832-78c5ab4512b1> lazy-loaded attributes: flavor,pci_requests {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 943.682091] env[62952]: DEBUG nova.network.neutron [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 943.774326] env[62952]: DEBUG nova.policy [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6967f8ebdc944dbfb2d19b5d7acdd3e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91fde3ceb94649f698459dfc174bf835', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 943.782033] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263456, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058425} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.782350] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.785964] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7ebe5ec-6389-4cce-8927-f7c299207bb0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.818235] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] e17b3d83-7cea-4a65-9457-7e57a6e7641a/e17b3d83-7cea-4a65-9457-7e57a6e7641a.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.820062] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21a7eb97-998d-420d-9284-dae318ae07b4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.846195] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 943.846195] env[62952]: value = "task-1263457" [ 943.846195] env[62952]: _type = "Task" [ 943.846195] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.856167] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263457, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.894535] env[62952]: DEBUG oslo_concurrency.lockutils [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "refresh_cache-20ea323b-8961-4981-8045-f9cb85815d99" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.894876] env[62952]: DEBUG nova.compute.manager [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Instance network_info: |[{"id": "8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855", "address": "fa:16:3e:65:72:4d", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ce9dcc3-e5", "ovs_interfaceid": "8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 943.895428] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:72:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '365ac5b1-6d83-4dfe-887f-60574d7f6124', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 943.903348] env[62952]: DEBUG oslo.service.loopingcall [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 943.903619] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 943.903855] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-36b3e29e-23e7-42a0-81b8-f931f759a9cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.919610] env[62952]: DEBUG nova.scheduler.client.report [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.927439] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 943.927439] env[62952]: value = "task-1263458" [ 943.927439] env[62952]: _type = "Task" [ 943.927439] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.935937] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263458, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.132580] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-91be5aa5-a3bc-49c4-8a22-cc99894849ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 944.132944] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-071e30d9-0c9c-4bb0-a91c-c077ff063a4a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.143830] env[62952]: DEBUG oslo_vmware.api [None req-91be5aa5-a3bc-49c4-8a22-cc99894849ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 944.143830] env[62952]: value = "task-1263459" [ 944.143830] env[62952]: _type = "Task" [ 944.143830] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.153531] env[62952]: DEBUG oslo_vmware.api [None req-91be5aa5-a3bc-49c4-8a22-cc99894849ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263459, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.360380] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263457, 'name': ReconfigVM_Task, 'duration_secs': 0.305877} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.360682] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Reconfigured VM instance instance-0000004a to attach disk [datastore1] e17b3d83-7cea-4a65-9457-7e57a6e7641a/e17b3d83-7cea-4a65-9457-7e57a6e7641a.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.361344] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cd87d7d5-7ab9-49db-8fc7-4d4451c66831 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.368200] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 944.368200] env[62952]: value = "task-1263460" [ 944.368200] env[62952]: _type = "Task" [ 944.368200] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.377376] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263460, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.424561] env[62952]: DEBUG oslo_concurrency.lockutils [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.958s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.427509] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.586s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.427509] env[62952]: DEBUG nova.objects.instance [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lazy-loading 'resources' on Instance uuid 47b57348-40d1-4735-9f5a-a7288f8f12ee {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.438271] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263458, 'name': CreateVM_Task, 'duration_secs': 0.474333} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.438497] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 944.439808] env[62952]: DEBUG oslo_concurrency.lockutils [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.439938] env[62952]: DEBUG oslo_concurrency.lockutils [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.440302] env[62952]: DEBUG oslo_concurrency.lockutils [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 944.441172] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-503d3670-9c68-4098-8898-862c6329b843 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.445873] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 944.445873] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52226bfa-b131-bc41-172c-f060e5b67f8d" [ 944.445873] env[62952]: _type = "Task" [ 944.445873] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.454191] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52226bfa-b131-bc41-172c-f060e5b67f8d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.459168] env[62952]: INFO nova.scheduler.client.report [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Deleted allocations for instance e5e7b2b5-34cf-4d37-87ff-1b57264a3516 [ 944.621262] env[62952]: DEBUG nova.compute.manager [req-2a08ee5a-d1cf-478e-ae8c-1c807db20c85 req-09968fd2-d8ba-470d-bee7-6fb59e5cc1b0 service nova] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Received event network-changed-8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 944.621470] env[62952]: DEBUG nova.compute.manager [req-2a08ee5a-d1cf-478e-ae8c-1c807db20c85 req-09968fd2-d8ba-470d-bee7-6fb59e5cc1b0 service nova] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Refreshing instance network info cache due to event network-changed-8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 944.621689] env[62952]: DEBUG oslo_concurrency.lockutils [req-2a08ee5a-d1cf-478e-ae8c-1c807db20c85 req-09968fd2-d8ba-470d-bee7-6fb59e5cc1b0 service nova] Acquiring lock "refresh_cache-20ea323b-8961-4981-8045-f9cb85815d99" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.621833] env[62952]: DEBUG oslo_concurrency.lockutils [req-2a08ee5a-d1cf-478e-ae8c-1c807db20c85 req-09968fd2-d8ba-470d-bee7-6fb59e5cc1b0 service nova] Acquired lock "refresh_cache-20ea323b-8961-4981-8045-f9cb85815d99" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.621992] env[62952]: DEBUG nova.network.neutron [req-2a08ee5a-d1cf-478e-ae8c-1c807db20c85 req-09968fd2-d8ba-470d-bee7-6fb59e5cc1b0 service nova] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Refreshing network info cache for port 8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 944.632328] env[62952]: DEBUG nova.compute.manager [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 944.634903] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de6e21a9-7f78-46dc-ae6c-eaf461de6acb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.652628] env[62952]: DEBUG oslo_vmware.api [None req-91be5aa5-a3bc-49c4-8a22-cc99894849ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263459, 'name': PowerOnVM_Task, 'duration_secs': 0.429575} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.652837] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-91be5aa5-a3bc-49c4-8a22-cc99894849ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 944.653063] env[62952]: DEBUG nova.compute.manager [None req-91be5aa5-a3bc-49c4-8a22-cc99894849ae tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 944.653832] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee744569-180a-4451-ab03-c63b965ec35d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.878819] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263460, 'name': Rename_Task, 'duration_secs': 0.139489} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.879165] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 944.879642] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c62120d9-5f53-40c8-ae5b-5b79805d5225 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.885830] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 944.885830] env[62952]: value = "task-1263461" [ 944.885830] env[62952]: _type = "Task" [ 944.885830] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.894336] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263461, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.956922] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52226bfa-b131-bc41-172c-f060e5b67f8d, 'name': SearchDatastore_Task, 'duration_secs': 0.009376} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.959650] env[62952]: DEBUG oslo_concurrency.lockutils [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.959887] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 944.960133] env[62952]: DEBUG oslo_concurrency.lockutils [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.960281] env[62952]: DEBUG oslo_concurrency.lockutils [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.960461] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 944.960953] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b0d67a28-ee01-4b26-bc79-22ce49f530b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.965622] env[62952]: DEBUG oslo_concurrency.lockutils [None req-41390a1b-3bea-459f-94d0-a74f1c2c4a56 tempest-MultipleCreateTestJSON-1978135428 tempest-MultipleCreateTestJSON-1978135428-project-member] Lock "e5e7b2b5-34cf-4d37-87ff-1b57264a3516" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.679s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.970197] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 944.970340] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 944.971724] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbb24193-793b-4bfd-a4f5-47929a46d379 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.981180] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 944.981180] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d3f189-a290-f708-e8de-37bea8b20cc5" [ 944.981180] env[62952]: _type = "Task" [ 944.981180] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.989650] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d3f189-a290-f708-e8de-37bea8b20cc5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.145938] env[62952]: INFO nova.compute.manager [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] instance snapshotting [ 945.149428] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6f8995c-1a53-4b9c-8806-9dfd65400ae9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.187926] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a320368e-2530-4911-a4cf-c89ea7688397 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.214149] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66975198-929b-437b-a96e-702dd36d7e2c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.222254] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-633e0711-5387-4b53-8bfd-9cb810771406 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.255284] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659546c7-6bad-4406-8d46-9f95e089ebde {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.263469] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8bfe26-90ef-4d3f-b9bb-1fdf423f6737 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.278476] env[62952]: DEBUG nova.compute.provider_tree [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.395711] env[62952]: DEBUG oslo_vmware.api [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263461, 'name': PowerOnVM_Task, 'duration_secs': 0.452651} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.396038] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 945.396228] env[62952]: INFO nova.compute.manager [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Took 8.28 seconds to spawn the instance on the hypervisor. [ 945.396413] env[62952]: DEBUG nova.compute.manager [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.397410] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-107a1e35-a9d7-484d-9661-e87f501c29da {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.493769] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d3f189-a290-f708-e8de-37bea8b20cc5, 'name': SearchDatastore_Task, 'duration_secs': 0.013457} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.494925] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d366c01-16f7-4fb4-911b-139e74a92d2c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.501573] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 945.501573] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d27d57-4762-145f-516b-e4eeb0daa2df" [ 945.501573] env[62952]: _type = "Task" [ 945.501573] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.510372] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d27d57-4762-145f-516b-e4eeb0daa2df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.596449] env[62952]: DEBUG nova.network.neutron [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Successfully updated port: bdfdd199-b7c7-47fd-bcd7-15760750128b {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 945.704980] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 945.705451] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b6ef6729-e1ba-42dd-af8f-a39ef126ac36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.720024] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 945.720024] env[62952]: value = "task-1263462" [ 945.720024] env[62952]: _type = "Task" [ 945.720024] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.728652] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263462, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.745645] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquiring lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.746835] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.775891] env[62952]: DEBUG nova.network.neutron [req-2a08ee5a-d1cf-478e-ae8c-1c807db20c85 req-09968fd2-d8ba-470d-bee7-6fb59e5cc1b0 service nova] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Updated VIF entry in instance network info cache for port 8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 945.776267] env[62952]: DEBUG nova.network.neutron [req-2a08ee5a-d1cf-478e-ae8c-1c807db20c85 req-09968fd2-d8ba-470d-bee7-6fb59e5cc1b0 service nova] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Updating instance_info_cache with network_info: [{"id": "8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855", "address": "fa:16:3e:65:72:4d", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ce9dcc3-e5", "ovs_interfaceid": "8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.782249] env[62952]: DEBUG nova.scheduler.client.report [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 945.919903] env[62952]: INFO nova.compute.manager [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Took 33.50 seconds to build instance. [ 946.013276] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d27d57-4762-145f-516b-e4eeb0daa2df, 'name': SearchDatastore_Task, 'duration_secs': 0.021193} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.013276] env[62952]: DEBUG oslo_concurrency.lockutils [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.013437] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 20ea323b-8961-4981-8045-f9cb85815d99/20ea323b-8961-4981-8045-f9cb85815d99.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 946.013756] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1139956f-9b58-493c-b237-1647e6a5b7f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.019822] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 946.019822] env[62952]: value = "task-1263463" [ 946.019822] env[62952]: _type = "Task" [ 946.019822] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.027730] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263463, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.098361] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.098519] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.098697] env[62952]: DEBUG nova.network.neutron [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 946.229418] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263462, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.248620] env[62952]: DEBUG nova.compute.manager [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 946.279292] env[62952]: DEBUG oslo_concurrency.lockutils [req-2a08ee5a-d1cf-478e-ae8c-1c807db20c85 req-09968fd2-d8ba-470d-bee7-6fb59e5cc1b0 service nova] Releasing lock "refresh_cache-20ea323b-8961-4981-8045-f9cb85815d99" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.294470] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.865s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.294997] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.894s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.296663] env[62952]: INFO nova.compute.claims [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 946.328196] env[62952]: INFO nova.scheduler.client.report [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Deleted allocations for instance 47b57348-40d1-4735-9f5a-a7288f8f12ee [ 946.423067] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e3537595-76bd-4f36-80e9-3a48b4ef32f9 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "e17b3d83-7cea-4a65-9457-7e57a6e7641a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.010s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.530467] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263463, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.646436] env[62952]: WARNING nova.network.neutron [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] 52af4564-da88-4cc8-b51b-07855da20a7c already exists in list: networks containing: ['52af4564-da88-4cc8-b51b-07855da20a7c']. ignoring it [ 946.731040] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263462, 'name': CreateSnapshot_Task, 'duration_secs': 0.693431} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.732250] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 946.735452] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aeda024-a95c-41ef-b94f-7b9036c89941 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.775741] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.824265] env[62952]: DEBUG nova.compute.manager [req-2eb48a00-12de-490f-9e88-382895692fe0 req-c1264cde-55f6-4ef0-87e1-9cf7ce33642b service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Received event network-vif-plugged-bdfdd199-b7c7-47fd-bcd7-15760750128b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 946.824502] env[62952]: DEBUG oslo_concurrency.lockutils [req-2eb48a00-12de-490f-9e88-382895692fe0 req-c1264cde-55f6-4ef0-87e1-9cf7ce33642b service nova] Acquiring lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.824754] env[62952]: DEBUG oslo_concurrency.lockutils [req-2eb48a00-12de-490f-9e88-382895692fe0 req-c1264cde-55f6-4ef0-87e1-9cf7ce33642b service nova] Lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.825420] env[62952]: DEBUG oslo_concurrency.lockutils [req-2eb48a00-12de-490f-9e88-382895692fe0 req-c1264cde-55f6-4ef0-87e1-9cf7ce33642b service nova] Lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.825420] env[62952]: DEBUG nova.compute.manager [req-2eb48a00-12de-490f-9e88-382895692fe0 req-c1264cde-55f6-4ef0-87e1-9cf7ce33642b service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] No waiting events found dispatching network-vif-plugged-bdfdd199-b7c7-47fd-bcd7-15760750128b {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 946.825420] env[62952]: WARNING nova.compute.manager [req-2eb48a00-12de-490f-9e88-382895692fe0 req-c1264cde-55f6-4ef0-87e1-9cf7ce33642b service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Received unexpected event network-vif-plugged-bdfdd199-b7c7-47fd-bcd7-15760750128b for instance with vm_state active and task_state None. [ 946.825420] env[62952]: DEBUG nova.compute.manager [req-2eb48a00-12de-490f-9e88-382895692fe0 req-c1264cde-55f6-4ef0-87e1-9cf7ce33642b service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Received event network-changed-bdfdd199-b7c7-47fd-bcd7-15760750128b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 946.825853] env[62952]: DEBUG nova.compute.manager [req-2eb48a00-12de-490f-9e88-382895692fe0 req-c1264cde-55f6-4ef0-87e1-9cf7ce33642b service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Refreshing instance network info cache due to event network-changed-bdfdd199-b7c7-47fd-bcd7-15760750128b. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 946.825853] env[62952]: DEBUG oslo_concurrency.lockutils [req-2eb48a00-12de-490f-9e88-382895692fe0 req-c1264cde-55f6-4ef0-87e1-9cf7ce33642b service nova] Acquiring lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.835833] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2d4ff17-c3a7-4888-8761-0c608441b1e4 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "47b57348-40d1-4735-9f5a-a7288f8f12ee" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.333s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.031552] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263463, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5874} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.032255] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 20ea323b-8961-4981-8045-f9cb85815d99/20ea323b-8961-4981-8045-f9cb85815d99.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 947.032637] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 947.033029] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d2c80557-2942-404d-a537-52a5c026c1ea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.042020] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 947.042020] env[62952]: value = "task-1263464" [ 947.042020] env[62952]: _type = "Task" [ 947.042020] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.049252] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263464, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.093925] env[62952]: DEBUG nova.network.neutron [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Updating instance_info_cache with network_info: [{"id": "b5b040c9-55a9-4929-b720-a1988f7492c9", "address": "fa:16:3e:0b:89:af", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5b040c9-55", "ovs_interfaceid": "b5b040c9-55a9-4929-b720-a1988f7492c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bdfdd199-b7c7-47fd-bcd7-15760750128b", "address": "fa:16:3e:08:b0:ff", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdfdd199-b7", "ovs_interfaceid": "bdfdd199-b7c7-47fd-bcd7-15760750128b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.254411] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 947.255461] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0a297891-4518-4c8d-8a8e-f7324a204a95 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.265450] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 947.265450] env[62952]: value = "task-1263465" [ 947.265450] env[62952]: _type = "Task" [ 947.265450] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.274276] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263465, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.557613] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263464, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.197165} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.557613] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 947.557613] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff7de0a-0da9-452b-bfc1-da487fb71647 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.582512] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 20ea323b-8961-4981-8045-f9cb85815d99/20ea323b-8961-4981-8045-f9cb85815d99.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 947.586703] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a4920bb-a7df-4b96-8976-9f34fc15ffbe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.602393] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.603293] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.603611] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.604028] env[62952]: DEBUG oslo_concurrency.lockutils [req-2eb48a00-12de-490f-9e88-382895692fe0 req-c1264cde-55f6-4ef0-87e1-9cf7ce33642b service nova] Acquired lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.604342] env[62952]: DEBUG nova.network.neutron [req-2eb48a00-12de-490f-9e88-382895692fe0 req-c1264cde-55f6-4ef0-87e1-9cf7ce33642b service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Refreshing network info cache for port bdfdd199-b7c7-47fd-bcd7-15760750128b {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 947.609023] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b5089c-7e08-4d0f-b5a3-a0599d286701 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.628041] env[62952]: DEBUG nova.virt.hardware [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 947.628531] env[62952]: DEBUG nova.virt.hardware [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 947.628819] env[62952]: DEBUG nova.virt.hardware [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 947.629185] env[62952]: DEBUG nova.virt.hardware [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 947.629468] env[62952]: DEBUG nova.virt.hardware [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 947.629782] env[62952]: DEBUG nova.virt.hardware [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 947.630126] env[62952]: DEBUG nova.virt.hardware [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 947.630400] env[62952]: DEBUG nova.virt.hardware [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 947.630674] env[62952]: DEBUG nova.virt.hardware [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 947.630958] env[62952]: DEBUG nova.virt.hardware [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 947.631780] env[62952]: DEBUG nova.virt.hardware [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 947.638305] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Reconfiguring VM to attach interface {{(pid=62952) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 947.639031] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 947.639031] env[62952]: value = "task-1263466" [ 947.639031] env[62952]: _type = "Task" [ 947.639031] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.640377] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25ae60ad-0906-4444-8c1c-c3230e719c47 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.653249] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f771eb57-ac9a-4c60-9c61-4cefbaba159e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.668739] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec62789-bd94-47f5-848c-9a07d98b7fd2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.671882] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263466, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.672185] env[62952]: DEBUG oslo_vmware.api [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 947.672185] env[62952]: value = "task-1263467" [ 947.672185] env[62952]: _type = "Task" [ 947.672185] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.701042] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea13f88-ad50-4b8a-8595-228918d10c93 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.709389] env[62952]: DEBUG oslo_vmware.api [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263467, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.716126] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b5d2fe-ffbe-4953-a53f-0ed81f44c707 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.737496] env[62952]: DEBUG nova.compute.provider_tree [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.744218] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "86ec7b3e-34d6-45e8-87b2-c8550f035ed0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.744461] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "86ec7b3e-34d6-45e8-87b2-c8550f035ed0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.744664] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "86ec7b3e-34d6-45e8-87b2-c8550f035ed0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.744882] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "86ec7b3e-34d6-45e8-87b2-c8550f035ed0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.745030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "86ec7b3e-34d6-45e8-87b2-c8550f035ed0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.747255] env[62952]: INFO nova.compute.manager [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Terminating instance [ 947.748929] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "refresh_cache-86ec7b3e-34d6-45e8-87b2-c8550f035ed0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.749101] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquired lock "refresh_cache-86ec7b3e-34d6-45e8-87b2-c8550f035ed0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.749272] env[62952]: DEBUG nova.network.neutron [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 947.776074] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263465, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.162321] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263466, 'name': ReconfigVM_Task, 'duration_secs': 0.438916} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.162695] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 20ea323b-8961-4981-8045-f9cb85815d99/20ea323b-8961-4981-8045-f9cb85815d99.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 948.163301] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-33cee46a-cf2e-4d01-a084-d902a16dd71a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.170574] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 948.170574] env[62952]: value = "task-1263468" [ 948.170574] env[62952]: _type = "Task" [ 948.170574] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.181698] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263468, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.185067] env[62952]: DEBUG oslo_vmware.api [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263467, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.242588] env[62952]: DEBUG nova.scheduler.client.report [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.276641] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263465, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.277615] env[62952]: DEBUG nova.network.neutron [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 948.352993] env[62952]: DEBUG nova.network.neutron [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.394520] env[62952]: DEBUG nova.network.neutron [req-2eb48a00-12de-490f-9e88-382895692fe0 req-c1264cde-55f6-4ef0-87e1-9cf7ce33642b service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Updated VIF entry in instance network info cache for port bdfdd199-b7c7-47fd-bcd7-15760750128b. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 948.395255] env[62952]: DEBUG nova.network.neutron [req-2eb48a00-12de-490f-9e88-382895692fe0 req-c1264cde-55f6-4ef0-87e1-9cf7ce33642b service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Updating instance_info_cache with network_info: [{"id": "b5b040c9-55a9-4929-b720-a1988f7492c9", "address": "fa:16:3e:0b:89:af", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5b040c9-55", "ovs_interfaceid": "b5b040c9-55a9-4929-b720-a1988f7492c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bdfdd199-b7c7-47fd-bcd7-15760750128b", "address": "fa:16:3e:08:b0:ff", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdfdd199-b7", "ovs_interfaceid": "bdfdd199-b7c7-47fd-bcd7-15760750128b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.686112] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263468, 'name': Rename_Task, 'duration_secs': 0.182381} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.689293] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 948.689575] env[62952]: DEBUG oslo_vmware.api [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263467, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.689884] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e5b5990-b5d9-47c5-bcdf-99aa448a6665 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.695435] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 948.695435] env[62952]: value = "task-1263469" [ 948.695435] env[62952]: _type = "Task" [ 948.695435] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.704497] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263469, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.749796] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.750355] env[62952]: DEBUG nova.compute.manager [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 948.753060] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.553s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.753253] env[62952]: DEBUG nova.objects.instance [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lazy-loading 'resources' on Instance uuid 7a73cb49-3e93-4fcc-b613-3808a8d5c64a {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.776380] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263465, 'name': CloneVM_Task} progress is 95%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.810568] env[62952]: INFO nova.compute.manager [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Rebuilding instance [ 948.856643] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Releasing lock "refresh_cache-86ec7b3e-34d6-45e8-87b2-c8550f035ed0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.857216] env[62952]: DEBUG nova.compute.manager [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 948.857440] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 948.860454] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c916d9-8602-4200-864e-43681ec8f7bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.863706] env[62952]: DEBUG nova.compute.manager [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.864517] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75445760-4be3-4632-a93c-a06c97164a3e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.875837] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 948.876290] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3efdaafd-f357-48c0-8a93-bf5959d758e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.882061] env[62952]: DEBUG oslo_vmware.api [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 948.882061] env[62952]: value = "task-1263470" [ 948.882061] env[62952]: _type = "Task" [ 948.882061] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.889992] env[62952]: DEBUG oslo_vmware.api [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263470, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.898734] env[62952]: DEBUG oslo_concurrency.lockutils [req-2eb48a00-12de-490f-9e88-382895692fe0 req-c1264cde-55f6-4ef0-87e1-9cf7ce33642b service nova] Releasing lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.186904] env[62952]: DEBUG oslo_vmware.api [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263467, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.204440] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263469, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.256545] env[62952]: DEBUG nova.compute.utils [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 949.260311] env[62952]: DEBUG nova.compute.manager [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 949.261023] env[62952]: DEBUG nova.network.neutron [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 949.277674] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263465, 'name': CloneVM_Task, 'duration_secs': 1.622682} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.280358] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Created linked-clone VM from snapshot [ 949.281337] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1aa42a2-d60a-4a13-a951-fd692477694b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.290114] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Uploading image d36ac993-3e17-46d7-a6d5-93c4255077dc {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 949.313703] env[62952]: DEBUG oslo_vmware.rw_handles [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 949.313703] env[62952]: value = "vm-271914" [ 949.313703] env[62952]: _type = "VirtualMachine" [ 949.313703] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 949.314056] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a8bcf437-b60d-4b31-8cad-6cd70d2e26dc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.319009] env[62952]: DEBUG nova.policy [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '469eae9db3d14e9f90988e3093514e69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff928791a3cf496cb8db2c2a6c47a6a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 949.321935] env[62952]: DEBUG oslo_vmware.rw_handles [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lease: (returnval){ [ 949.321935] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b62f6d-4843-0971-0c49-9810227f448b" [ 949.321935] env[62952]: _type = "HttpNfcLease" [ 949.321935] env[62952]: } obtained for exporting VM: (result){ [ 949.321935] env[62952]: value = "vm-271914" [ 949.321935] env[62952]: _type = "VirtualMachine" [ 949.321935] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 949.322215] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the lease: (returnval){ [ 949.322215] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b62f6d-4843-0971-0c49-9810227f448b" [ 949.322215] env[62952]: _type = "HttpNfcLease" [ 949.322215] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 949.328733] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 949.328733] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b62f6d-4843-0971-0c49-9810227f448b" [ 949.328733] env[62952]: _type = "HttpNfcLease" [ 949.328733] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 949.378168] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 949.378410] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be6485b6-5b17-4e18-bb9c-37336a91049a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.389801] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 949.389801] env[62952]: value = "task-1263472" [ 949.389801] env[62952]: _type = "Task" [ 949.389801] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.397290] env[62952]: DEBUG oslo_vmware.api [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263470, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.400648] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263472, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.527861] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-611de7a2-392a-4eb9-809d-9c1464dd3909 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.535662] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59598a5e-dc63-4e80-914c-30485715333f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.568298] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd563d36-33bd-4352-8550-cfc3b7bc956e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.576019] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0897cf27-276e-487c-9468-cff71e901ed0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.589657] env[62952]: DEBUG nova.compute.provider_tree [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.659443] env[62952]: DEBUG nova.network.neutron [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Successfully created port: b708a98e-0fd2-4730-8307-8731609c4aa3 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 949.688134] env[62952]: DEBUG oslo_vmware.api [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263467, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.705617] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263469, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.761128] env[62952]: DEBUG nova.compute.manager [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 949.832067] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 949.832067] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b62f6d-4843-0971-0c49-9810227f448b" [ 949.832067] env[62952]: _type = "HttpNfcLease" [ 949.832067] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 949.832067] env[62952]: DEBUG oslo_vmware.rw_handles [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 949.832067] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b62f6d-4843-0971-0c49-9810227f448b" [ 949.832067] env[62952]: _type = "HttpNfcLease" [ 949.832067] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 949.832468] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c14d73-16e8-4b6f-9aec-e108b02f540c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.839670] env[62952]: DEBUG oslo_vmware.rw_handles [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5241db49-b1ed-9adb-1eb8-ee8a1671af95/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 949.839860] env[62952]: DEBUG oslo_vmware.rw_handles [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5241db49-b1ed-9adb-1eb8-ee8a1671af95/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 949.914160] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263472, 'name': PowerOffVM_Task, 'duration_secs': 0.262156} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.917602] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 949.917740] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 949.918042] env[62952]: DEBUG oslo_vmware.api [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263470, 'name': PowerOffVM_Task, 'duration_secs': 0.879587} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.919222] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb4955c-20a2-4358-8e23-7fa33d814227 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.921735] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 949.922043] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 949.922199] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bc35307f-8bb2-42e9-8892-0b59ef58fb79 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.929812] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 949.930120] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-35892bf2-4afb-4c35-a7ed-ec9441b43d6c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.940779] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-793fe087-fd00-47c8-846c-2b7f69d01827 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.951558] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 949.951801] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 949.952070] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Deleting the datastore file [datastore2] 86ec7b3e-34d6-45e8-87b2-c8550f035ed0 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 949.952468] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fcd76d4d-af9c-4c89-baa1-49a22065ee34 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.961689] env[62952]: DEBUG oslo_vmware.api [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for the task: (returnval){ [ 949.961689] env[62952]: value = "task-1263475" [ 949.961689] env[62952]: _type = "Task" [ 949.961689] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.977859] env[62952]: DEBUG oslo_vmware.api [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263475, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.007044] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 950.007044] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 950.007574] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleting the datastore file [datastore1] e17b3d83-7cea-4a65-9457-7e57a6e7641a {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.007574] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-51b5168b-0fe0-441c-a6ac-7df29bc3ff94 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.014796] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 950.014796] env[62952]: value = "task-1263476" [ 950.014796] env[62952]: _type = "Task" [ 950.014796] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.023843] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263476, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.093428] env[62952]: DEBUG nova.scheduler.client.report [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.187931] env[62952]: DEBUG oslo_vmware.api [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263467, 'name': ReconfigVM_Task, 'duration_secs': 2.091185} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.188495] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.188719] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Reconfigured VM to attach interface {{(pid=62952) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 950.207073] env[62952]: DEBUG oslo_vmware.api [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263469, 'name': PowerOnVM_Task, 'duration_secs': 1.39947} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.207430] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 950.207666] env[62952]: INFO nova.compute.manager [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Took 8.62 seconds to spawn the instance on the hypervisor. [ 950.207872] env[62952]: DEBUG nova.compute.manager [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 950.209366] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f0163b-8534-49da-9fc4-4030e02f86e8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.476598] env[62952]: DEBUG oslo_vmware.api [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Task: {'id': task-1263475, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102082} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.476882] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.477118] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 950.477247] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 950.477427] env[62952]: INFO nova.compute.manager [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Took 1.62 seconds to destroy the instance on the hypervisor. [ 950.477689] env[62952]: DEBUG oslo.service.loopingcall [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.477845] env[62952]: DEBUG nova.compute.manager [-] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 950.477941] env[62952]: DEBUG nova.network.neutron [-] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 950.495050] env[62952]: DEBUG nova.network.neutron [-] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 950.524960] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263476, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150138} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.525413] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.525812] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 950.526165] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 950.599026] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.846s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.601823] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.020s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.603838] env[62952]: INFO nova.compute.claims [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.629855] env[62952]: INFO nova.scheduler.client.report [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Deleted allocations for instance 7a73cb49-3e93-4fcc-b613-3808a8d5c64a [ 950.693121] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0ba79a73-d97e-4689-8a00-76ba9b479cc3 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "interface-0533ee81-b69d-4f4e-9832-78c5ab4512b1-bdfdd199-b7c7-47fd-bcd7-15760750128b" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.220s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.729212] env[62952]: INFO nova.compute.manager [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Took 33.76 seconds to build instance. [ 950.769249] env[62952]: DEBUG nova.compute.manager [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 950.876837] env[62952]: DEBUG nova.virt.hardware [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 950.877301] env[62952]: DEBUG nova.virt.hardware [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 950.877554] env[62952]: DEBUG nova.virt.hardware [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 950.877814] env[62952]: DEBUG nova.virt.hardware [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 950.878113] env[62952]: DEBUG nova.virt.hardware [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 950.878301] env[62952]: DEBUG nova.virt.hardware [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 950.878552] env[62952]: DEBUG nova.virt.hardware [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 950.878692] env[62952]: DEBUG nova.virt.hardware [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 950.878879] env[62952]: DEBUG nova.virt.hardware [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 950.879062] env[62952]: DEBUG nova.virt.hardware [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 950.879240] env[62952]: DEBUG nova.virt.hardware [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.880573] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d3334a-5d18-4ade-8e1a-2885b8f9f02b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.889381] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae83c856-f3d5-4909-8ee9-cc68c09abc84 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.997817] env[62952]: DEBUG nova.network.neutron [-] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.140078] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c739d4b-6f5a-4a99-969f-641c1234cf9f tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "7a73cb49-3e93-4fcc-b613-3808a8d5c64a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.109s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.231516] env[62952]: DEBUG oslo_concurrency.lockutils [None req-648c335f-a7b6-4320-ab0a-67d0027d88b3 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "20ea323b-8961-4981-8045-f9cb85815d99" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.271s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.327946] env[62952]: DEBUG nova.network.neutron [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Successfully updated port: b708a98e-0fd2-4730-8307-8731609c4aa3 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 951.462678] env[62952]: DEBUG nova.compute.manager [req-61675d2e-eddd-4c3f-9e9b-dc74bd298df3 req-916166a9-7cda-416d-b6fc-bd9733d53e24 service nova] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Received event network-vif-plugged-b708a98e-0fd2-4730-8307-8731609c4aa3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 951.463114] env[62952]: DEBUG oslo_concurrency.lockutils [req-61675d2e-eddd-4c3f-9e9b-dc74bd298df3 req-916166a9-7cda-416d-b6fc-bd9733d53e24 service nova] Acquiring lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.463592] env[62952]: DEBUG oslo_concurrency.lockutils [req-61675d2e-eddd-4c3f-9e9b-dc74bd298df3 req-916166a9-7cda-416d-b6fc-bd9733d53e24 service nova] Lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.463835] env[62952]: DEBUG oslo_concurrency.lockutils [req-61675d2e-eddd-4c3f-9e9b-dc74bd298df3 req-916166a9-7cda-416d-b6fc-bd9733d53e24 service nova] Lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.464096] env[62952]: DEBUG nova.compute.manager [req-61675d2e-eddd-4c3f-9e9b-dc74bd298df3 req-916166a9-7cda-416d-b6fc-bd9733d53e24 service nova] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] No waiting events found dispatching network-vif-plugged-b708a98e-0fd2-4730-8307-8731609c4aa3 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 951.464393] env[62952]: WARNING nova.compute.manager [req-61675d2e-eddd-4c3f-9e9b-dc74bd298df3 req-916166a9-7cda-416d-b6fc-bd9733d53e24 service nova] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Received unexpected event network-vif-plugged-b708a98e-0fd2-4730-8307-8731609c4aa3 for instance with vm_state building and task_state spawning. [ 951.501572] env[62952]: INFO nova.compute.manager [-] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Took 1.02 seconds to deallocate network for instance. [ 951.557981] env[62952]: DEBUG nova.virt.hardware [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 951.558515] env[62952]: DEBUG nova.virt.hardware [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 951.558640] env[62952]: DEBUG nova.virt.hardware [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.558886] env[62952]: DEBUG nova.virt.hardware [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 951.559097] env[62952]: DEBUG nova.virt.hardware [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.559445] env[62952]: DEBUG nova.virt.hardware [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 951.559685] env[62952]: DEBUG nova.virt.hardware [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 951.559858] env[62952]: DEBUG nova.virt.hardware [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 951.560050] env[62952]: DEBUG nova.virt.hardware [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 951.560233] env[62952]: DEBUG nova.virt.hardware [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 951.560419] env[62952]: DEBUG nova.virt.hardware [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 951.561417] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b0ca822-6d60-4afc-936a-e1338d5c1921 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.570061] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0729e5f-bf77-4ac9-ace0-37914bea7e4f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.585501] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:17:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5ad8894f-e240-4013-8272-4e79daea0751', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd1e678fd-723e-413a-a8e8-44e88b909665', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.594097] env[62952]: DEBUG oslo.service.loopingcall [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.594596] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 951.594596] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4dbe5d33-df19-407d-874a-1fb3e3a51dcf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.620372] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.620372] env[62952]: value = "task-1263477" [ 951.620372] env[62952]: _type = "Task" [ 951.620372] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.630728] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263477, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.705451] env[62952]: DEBUG oslo_concurrency.lockutils [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "5f1eca6f-4cc4-483e-b72a-b479378277fe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.705451] env[62952]: DEBUG oslo_concurrency.lockutils [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "5f1eca6f-4cc4-483e-b72a-b479378277fe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.705621] env[62952]: DEBUG oslo_concurrency.lockutils [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "5f1eca6f-4cc4-483e-b72a-b479378277fe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.705957] env[62952]: DEBUG oslo_concurrency.lockutils [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "5f1eca6f-4cc4-483e-b72a-b479378277fe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.706311] env[62952]: DEBUG oslo_concurrency.lockutils [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "5f1eca6f-4cc4-483e-b72a-b479378277fe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.709668] env[62952]: INFO nova.compute.manager [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Terminating instance [ 951.712084] env[62952]: DEBUG nova.compute.manager [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 951.712444] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 951.716665] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b2c042-7679-4929-80d9-7f9eff662084 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.725642] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 951.726774] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-abe91d48-9d44-4124-9200-5787ddbf0c3b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.734551] env[62952]: DEBUG oslo_vmware.api [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 951.734551] env[62952]: value = "task-1263478" [ 951.734551] env[62952]: _type = "Task" [ 951.734551] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.743639] env[62952]: DEBUG oslo_vmware.api [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263478, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.837474] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "refresh_cache-b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.837474] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "refresh_cache-b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.837474] env[62952]: DEBUG nova.network.neutron [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 951.894969] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f9ca7e1-ee1a-4c23-818e-8a41e3598b6a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.904056] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d205b0-0cf5-4cd5-94df-8f40845620bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.937158] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d496dfa-afad-44c6-88fb-ceb82de88fc3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.949020] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de7cbf23-a805-4085-99c2-f3fd8839215c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.965036] env[62952]: DEBUG nova.compute.provider_tree [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.007951] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.901363] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "20ea323b-8961-4981-8045-f9cb85815d99" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.901746] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "20ea323b-8961-4981-8045-f9cb85815d99" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.901809] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "20ea323b-8961-4981-8045-f9cb85815d99-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.901973] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "20ea323b-8961-4981-8045-f9cb85815d99-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.902161] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "20ea323b-8961-4981-8045-f9cb85815d99-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.904508] env[62952]: DEBUG nova.scheduler.client.report [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.907776] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263477, 'name': CreateVM_Task, 'duration_secs': 0.401673} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.911120] env[62952]: INFO nova.compute.manager [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Terminating instance [ 952.917210] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 952.918643] env[62952]: DEBUG nova.compute.manager [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 952.918643] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 952.918643] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "c7d08234-fcef-4cdb-939f-05ea9e8b08ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.918840] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "c7d08234-fcef-4cdb-939f-05ea9e8b08ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.920518] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.920801] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.921188] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 952.921766] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e5b0b2-f627-4ca3-8c61-7c0fc8ccafe0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.926162] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82b77e7e-9206-451b-9ec4-e0b14edd711d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.932682] env[62952]: DEBUG oslo_vmware.api [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263478, 'name': PowerOffVM_Task, 'duration_secs': 0.241231} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.935885] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 952.936119] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 952.936657] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 952.936657] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d40e6c-3ff3-f1f5-250c-fcfa4224c909" [ 952.936657] env[62952]: _type = "Task" [ 952.936657] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.936891] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 952.937127] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ee9cf06e-5ea2-4b9d-8b92-ae7dd63b4028 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.938739] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f5dcec4-0d3b-442a-bca2-37f9ecce006c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.948984] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d40e6c-3ff3-f1f5-250c-fcfa4224c909, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.950353] env[62952]: DEBUG oslo_vmware.api [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 952.950353] env[62952]: value = "task-1263479" [ 952.950353] env[62952]: _type = "Task" [ 952.950353] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.958989] env[62952]: DEBUG oslo_vmware.api [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263479, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.961133] env[62952]: DEBUG nova.network.neutron [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 953.010394] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 953.010726] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 953.010849] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Deleting the datastore file [datastore1] 5f1eca6f-4cc4-483e-b72a-b479378277fe {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 953.011207] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e01ce4fe-71c2-40e3-abaf-eb1518237432 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.017847] env[62952]: DEBUG oslo_vmware.api [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 953.017847] env[62952]: value = "task-1263481" [ 953.017847] env[62952]: _type = "Task" [ 953.017847] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.026473] env[62952]: DEBUG oslo_vmware.api [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263481, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.140679] env[62952]: DEBUG nova.network.neutron [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Updating instance_info_cache with network_info: [{"id": "b708a98e-0fd2-4730-8307-8731609c4aa3", "address": "fa:16:3e:a2:71:17", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb708a98e-0f", "ovs_interfaceid": "b708a98e-0fd2-4730-8307-8731609c4aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.409965] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.808s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.410997] env[62952]: DEBUG nova.compute.manager [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 953.413115] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.686s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.413351] env[62952]: DEBUG nova.objects.instance [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lazy-loading 'resources' on Instance uuid c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.418350] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "interface-0533ee81-b69d-4f4e-9832-78c5ab4512b1-bdfdd199-b7c7-47fd-bcd7-15760750128b" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.418350] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "interface-0533ee81-b69d-4f4e-9832-78c5ab4512b1-bdfdd199-b7c7-47fd-bcd7-15760750128b" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.425045] env[62952]: DEBUG nova.compute.manager [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 953.449207] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d40e6c-3ff3-f1f5-250c-fcfa4224c909, 'name': SearchDatastore_Task, 'duration_secs': 0.021299} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.449843] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.449843] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 953.450072] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.450127] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.450296] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.450600] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-130c999a-5a9d-46dd-825b-f65b8cbc8ea3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.461695] env[62952]: DEBUG oslo_vmware.api [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263479, 'name': PowerOffVM_Task, 'duration_secs': 0.22529} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.462408] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 953.462582] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 953.463329] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25e37e8e-b612-4ae5-b8d7-e15d526c323b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.464854] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.465567] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 953.466030] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31f7760e-6f1f-4c9c-8431-4c7734d17b54 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.473778] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 953.473778] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d54e30-9a9e-5fe9-3b17-8610dae688a9" [ 953.473778] env[62952]: _type = "Task" [ 953.473778] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.482079] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d54e30-9a9e-5fe9-3b17-8610dae688a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.489283] env[62952]: DEBUG nova.compute.manager [req-51755b1c-e418-4a48-ab2a-a2aeedd9dccd req-84f8a777-1c65-4510-a6a7-fdc5658078cc service nova] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Received event network-changed-b708a98e-0fd2-4730-8307-8731609c4aa3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 953.489401] env[62952]: DEBUG nova.compute.manager [req-51755b1c-e418-4a48-ab2a-a2aeedd9dccd req-84f8a777-1c65-4510-a6a7-fdc5658078cc service nova] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Refreshing instance network info cache due to event network-changed-b708a98e-0fd2-4730-8307-8731609c4aa3. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 953.489593] env[62952]: DEBUG oslo_concurrency.lockutils [req-51755b1c-e418-4a48-ab2a-a2aeedd9dccd req-84f8a777-1c65-4510-a6a7-fdc5658078cc service nova] Acquiring lock "refresh_cache-b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.527997] env[62952]: DEBUG oslo_vmware.api [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263481, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198217} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.528284] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.528471] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 953.528649] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 953.528822] env[62952]: INFO nova.compute.manager [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Took 1.82 seconds to destroy the instance on the hypervisor. [ 953.529094] env[62952]: DEBUG oslo.service.loopingcall [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.529294] env[62952]: DEBUG nova.compute.manager [-] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 953.529392] env[62952]: DEBUG nova.network.neutron [-] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 953.550613] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 953.550776] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 953.550954] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleting the datastore file [datastore1] 20ea323b-8961-4981-8045-f9cb85815d99 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 953.551249] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-276499f9-2f6d-41f0-83e6-b4bdba7ae291 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.559915] env[62952]: DEBUG oslo_vmware.api [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 953.559915] env[62952]: value = "task-1263483" [ 953.559915] env[62952]: _type = "Task" [ 953.559915] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.570266] env[62952]: DEBUG oslo_vmware.api [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263483, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.643290] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "refresh_cache-b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.643757] env[62952]: DEBUG nova.compute.manager [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Instance network_info: |[{"id": "b708a98e-0fd2-4730-8307-8731609c4aa3", "address": "fa:16:3e:a2:71:17", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb708a98e-0f", "ovs_interfaceid": "b708a98e-0fd2-4730-8307-8731609c4aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 953.644145] env[62952]: DEBUG oslo_concurrency.lockutils [req-51755b1c-e418-4a48-ab2a-a2aeedd9dccd req-84f8a777-1c65-4510-a6a7-fdc5658078cc service nova] Acquired lock "refresh_cache-b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.644367] env[62952]: DEBUG nova.network.neutron [req-51755b1c-e418-4a48-ab2a-a2aeedd9dccd req-84f8a777-1c65-4510-a6a7-fdc5658078cc service nova] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Refreshing network info cache for port b708a98e-0fd2-4730-8307-8731609c4aa3 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 953.649026] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:71:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b708a98e-0fd2-4730-8307-8731609c4aa3', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 953.653712] env[62952]: DEBUG oslo.service.loopingcall [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.654775] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 953.655025] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-54875e81-a893-4549-af3e-aa7d86d4a189 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.675408] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 953.675408] env[62952]: value = "task-1263484" [ 953.675408] env[62952]: _type = "Task" [ 953.675408] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.687355] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263484, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.917042] env[62952]: DEBUG nova.compute.utils [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 953.920962] env[62952]: DEBUG nova.compute.manager [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 953.921187] env[62952]: DEBUG nova.network.neutron [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 953.923071] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.923274] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.924074] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9748c54-63f7-4272-81a7-bf41a317fdd0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.945680] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-155b9958-216b-4db8-9940-c5e4841b24fa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.976316] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Reconfiguring VM to detach interface {{(pid=62952) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 953.977811] env[62952]: DEBUG nova.policy [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '23bdda22bcf74a1ea9d681512d663aeb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '091cdeb6da48492bb02f93822a45c9bf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 953.979974] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.982314] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-884dd875-602b-41e9-8a5d-5f87bada2121 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.007105] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d54e30-9a9e-5fe9-3b17-8610dae688a9, 'name': SearchDatastore_Task, 'duration_secs': 0.012226} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.011951] env[62952]: DEBUG oslo_vmware.api [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 954.011951] env[62952]: value = "task-1263485" [ 954.011951] env[62952]: _type = "Task" [ 954.011951] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.012812] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1ceac47-bd24-4774-8ddb-a6c1d6da4441 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.021997] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 954.021997] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]527c17a1-6a1f-32a2-bef5-261847f77f63" [ 954.021997] env[62952]: _type = "Task" [ 954.021997] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.027847] env[62952]: DEBUG oslo_vmware.api [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263485, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.036150] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]527c17a1-6a1f-32a2-bef5-261847f77f63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.071811] env[62952]: DEBUG oslo_vmware.api [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263483, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.268254} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.074390] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 954.074593] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 954.074774] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 954.074950] env[62952]: INFO nova.compute.manager [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Took 1.16 seconds to destroy the instance on the hypervisor. [ 954.075221] env[62952]: DEBUG oslo.service.loopingcall [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.075788] env[62952]: DEBUG nova.compute.manager [-] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 954.076189] env[62952]: DEBUG nova.network.neutron [-] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 954.186559] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263484, 'name': CreateVM_Task, 'duration_secs': 0.484785} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.186734] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 954.187458] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.187653] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.188310] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 954.188310] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a572de5d-aa63-438c-af3c-b3ad69b84e8b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.192998] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 954.192998] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520df1f5-d904-e794-5683-b60aafbd753b" [ 954.192998] env[62952]: _type = "Task" [ 954.192998] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.207827] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520df1f5-d904-e794-5683-b60aafbd753b, 'name': SearchDatastore_Task, 'duration_secs': 0.008988} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.210764] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.211117] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 954.211342] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.250248] env[62952]: DEBUG nova.network.neutron [-] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.257542] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a441f97-ec2c-4dd5-93c9-37fcf8213034 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.266431] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4f1d48-f7da-410a-9d62-33c4d2f86b12 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.303649] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1680aedb-a987-4d19-8a0d-b5efc2dc4f26 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.311911] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce95f12f-ba64-4564-a8d1-bde80feb6f1a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.329031] env[62952]: DEBUG nova.compute.provider_tree [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.395430] env[62952]: DEBUG nova.compute.manager [req-9fe2c32d-c117-43b2-a97b-9ea09956c5c8 req-6c81a48a-f160-4858-8003-feef92fb9781 service nova] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Received event network-vif-deleted-8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 954.395576] env[62952]: INFO nova.compute.manager [req-9fe2c32d-c117-43b2-a97b-9ea09956c5c8 req-6c81a48a-f160-4858-8003-feef92fb9781 service nova] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Neutron deleted interface 8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855; detaching it from the instance and deleting it from the info cache [ 954.395750] env[62952]: DEBUG nova.network.neutron [req-9fe2c32d-c117-43b2-a97b-9ea09956c5c8 req-6c81a48a-f160-4858-8003-feef92fb9781 service nova] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.422025] env[62952]: DEBUG nova.compute.manager [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 954.526984] env[62952]: DEBUG oslo_vmware.api [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263485, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.537674] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]527c17a1-6a1f-32a2-bef5-261847f77f63, 'name': SearchDatastore_Task, 'duration_secs': 0.013272} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.537936] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.538213] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] e17b3d83-7cea-4a65-9457-7e57a6e7641a/e17b3d83-7cea-4a65-9457-7e57a6e7641a.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 954.538489] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.538677] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 954.538886] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c56c275b-edd0-4764-9c41-889d9cb8df34 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.540711] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1fc4c653-0402-4865-b9d1-e7f1f68c7cad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.547432] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 954.547432] env[62952]: value = "task-1263486" [ 954.547432] env[62952]: _type = "Task" [ 954.547432] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.551792] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 954.551971] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 954.552966] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-686dcffb-a7c6-4600-bee9-4f150fd1e398 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.558532] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263486, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.561377] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 954.561377] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52fb19a0-84f7-72da-e9ff-859905fd839f" [ 954.561377] env[62952]: _type = "Task" [ 954.561377] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.569066] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52fb19a0-84f7-72da-e9ff-859905fd839f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.755658] env[62952]: INFO nova.compute.manager [-] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Took 1.22 seconds to deallocate network for instance. [ 954.816430] env[62952]: DEBUG nova.network.neutron [req-51755b1c-e418-4a48-ab2a-a2aeedd9dccd req-84f8a777-1c65-4510-a6a7-fdc5658078cc service nova] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Updated VIF entry in instance network info cache for port b708a98e-0fd2-4730-8307-8731609c4aa3. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 954.816912] env[62952]: DEBUG nova.network.neutron [req-51755b1c-e418-4a48-ab2a-a2aeedd9dccd req-84f8a777-1c65-4510-a6a7-fdc5658078cc service nova] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Updating instance_info_cache with network_info: [{"id": "b708a98e-0fd2-4730-8307-8731609c4aa3", "address": "fa:16:3e:a2:71:17", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb708a98e-0f", "ovs_interfaceid": "b708a98e-0fd2-4730-8307-8731609c4aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.819200] env[62952]: DEBUG nova.network.neutron [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Successfully created port: d93fe914-641a-4750-a765-181c7012b252 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 954.828943] env[62952]: DEBUG nova.scheduler.client.report [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.870239] env[62952]: DEBUG nova.network.neutron [-] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.901019] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb51c40b-e82b-41bc-bf31-42c8bce02a0c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.912031] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e50499a7-e570-43c1-8311-19354dac5095 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.943840] env[62952]: DEBUG nova.compute.manager [req-9fe2c32d-c117-43b2-a97b-9ea09956c5c8 req-6c81a48a-f160-4858-8003-feef92fb9781 service nova] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Detach interface failed, port_id=8ce9dcc3-e5fe-4ecb-baaa-29d21a78f855, reason: Instance 20ea323b-8961-4981-8045-f9cb85815d99 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 955.025793] env[62952]: DEBUG oslo_vmware.api [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263485, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.057779] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263486, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.444239} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.058272] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] e17b3d83-7cea-4a65-9457-7e57a6e7641a/e17b3d83-7cea-4a65-9457-7e57a6e7641a.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 955.058692] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 955.059121] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ae78fd0e-3f1a-47c9-93c2-331855990e8c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.069434] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 955.069434] env[62952]: value = "task-1263487" [ 955.069434] env[62952]: _type = "Task" [ 955.069434] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.075093] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52fb19a0-84f7-72da-e9ff-859905fd839f, 'name': SearchDatastore_Task, 'duration_secs': 0.007752} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.077119] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51c70a79-590c-41f1-82f3-6d13591183d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.082567] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263487, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.085785] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 955.085785] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5215c8d5-886a-485c-7403-1758c07a6db6" [ 955.085785] env[62952]: _type = "Task" [ 955.085785] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.093136] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5215c8d5-886a-485c-7403-1758c07a6db6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.265010] env[62952]: DEBUG oslo_concurrency.lockutils [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.322968] env[62952]: DEBUG oslo_concurrency.lockutils [req-51755b1c-e418-4a48-ab2a-a2aeedd9dccd req-84f8a777-1c65-4510-a6a7-fdc5658078cc service nova] Releasing lock "refresh_cache-b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.335993] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.923s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.341254] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.645s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.341477] env[62952]: INFO nova.compute.claims [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 955.362503] env[62952]: INFO nova.scheduler.client.report [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Deleted allocations for instance c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb [ 955.373606] env[62952]: INFO nova.compute.manager [-] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Took 1.30 seconds to deallocate network for instance. [ 955.430625] env[62952]: DEBUG nova.compute.manager [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 955.456042] env[62952]: DEBUG nova.virt.hardware [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 955.456370] env[62952]: DEBUG nova.virt.hardware [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 955.456540] env[62952]: DEBUG nova.virt.hardware [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 955.456749] env[62952]: DEBUG nova.virt.hardware [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 955.456900] env[62952]: DEBUG nova.virt.hardware [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 955.457092] env[62952]: DEBUG nova.virt.hardware [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 955.457337] env[62952]: DEBUG nova.virt.hardware [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 955.457506] env[62952]: DEBUG nova.virt.hardware [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 955.457694] env[62952]: DEBUG nova.virt.hardware [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 955.457862] env[62952]: DEBUG nova.virt.hardware [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 955.458072] env[62952]: DEBUG nova.virt.hardware [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 955.458994] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be8ab968-f6bb-4a2e-97ef-44d56962915f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.467167] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd717b6-8c59-415d-9daf-3a141b72f036 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.517370] env[62952]: DEBUG nova.compute.manager [req-7f586730-319d-4b2c-94e1-0efd876edf4f req-f8eda153-d988-4d72-af7a-a171696d9a0a service nova] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Received event network-vif-deleted-b4e55382-c684-4ae1-afc7-0831ef60ac7e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 955.528542] env[62952]: DEBUG oslo_vmware.api [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263485, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.577983] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263487, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064964} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.578302] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 955.579124] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d0b137-174b-4664-8f61-6b468aaf24e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.602573] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] e17b3d83-7cea-4a65-9457-7e57a6e7641a/e17b3d83-7cea-4a65-9457-7e57a6e7641a.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.605988] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d9e4943-224d-4873-9482-e3c9fba019d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.627047] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5215c8d5-886a-485c-7403-1758c07a6db6, 'name': SearchDatastore_Task, 'duration_secs': 0.016816} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.628279] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.628551] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e/b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 955.628880] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 955.628880] env[62952]: value = "task-1263488" [ 955.628880] env[62952]: _type = "Task" [ 955.628880] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.629095] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7b53ba72-502f-4678-9b59-260e7133d199 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.639653] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263488, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.640930] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 955.640930] env[62952]: value = "task-1263489" [ 955.640930] env[62952]: _type = "Task" [ 955.640930] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.649517] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263489, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.871143] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e208271c-5d2d-4ea9-b7ba-1c650a6a4b34 tempest-ListServerFiltersTestJSON-1379770526 tempest-ListServerFiltersTestJSON-1379770526-project-member] Lock "c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.353s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.881451] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.030862] env[62952]: DEBUG oslo_vmware.api [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263485, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.141168] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263488, 'name': ReconfigVM_Task, 'duration_secs': 0.456279} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.141650] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Reconfigured VM instance instance-0000004a to attach disk [datastore2] e17b3d83-7cea-4a65-9457-7e57a6e7641a/e17b3d83-7cea-4a65-9457-7e57a6e7641a.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.145159] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d61d4b8b-b121-4e78-9717-af71e257af15 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.152881] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263489, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447118} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.154261] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e/b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 956.154499] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 956.154845] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 956.154845] env[62952]: value = "task-1263490" [ 956.154845] env[62952]: _type = "Task" [ 956.154845] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.155080] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c60616c0-be55-4b6f-be2b-841577d4d0b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.165580] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263490, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.166955] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 956.166955] env[62952]: value = "task-1263491" [ 956.166955] env[62952]: _type = "Task" [ 956.166955] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.175412] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263491, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.530905] env[62952]: DEBUG oslo_vmware.api [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263485, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.581087] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e1078f-ac6a-40c5-b38d-301b0a7a46f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.590714] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a3d44ed-dc15-4493-8f38-497b7c4f73cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.624382] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745647f1-87fa-419a-8a57-5acbdb843a51 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.632911] env[62952]: DEBUG nova.network.neutron [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Successfully updated port: d93fe914-641a-4750-a765-181c7012b252 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 956.634497] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10405bb7-ad0e-4d1c-a6a8-8fa0179f0725 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.654038] env[62952]: DEBUG nova.compute.provider_tree [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.668532] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263490, 'name': Rename_Task, 'duration_secs': 0.21783} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.668532] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 956.671878] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f4e0f902-6a6b-416d-9c08-9ae37c4f59bc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.679813] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263491, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.123788} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.681400] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 956.681850] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 956.681850] env[62952]: value = "task-1263492" [ 956.681850] env[62952]: _type = "Task" [ 956.681850] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.682713] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69eb6b98-7523-4b98-bc15-bbce0440786e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.712220] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e/b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.715852] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89ae7ae3-27e2-4d82-9515-4a7c25d43a6f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.731140] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263492, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.739076] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 956.739076] env[62952]: value = "task-1263493" [ 956.739076] env[62952]: _type = "Task" [ 956.739076] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.747576] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263493, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.028701] env[62952]: DEBUG oslo_vmware.api [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263485, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.141934] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "refresh_cache-dbca4369-1ed3-493a-9847-9fa1e4293475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.142291] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired lock "refresh_cache-dbca4369-1ed3-493a-9847-9fa1e4293475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.142291] env[62952]: DEBUG nova.network.neutron [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 957.156484] env[62952]: DEBUG nova.scheduler.client.report [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 957.201427] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263492, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.252167] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263493, 'name': ReconfigVM_Task, 'duration_secs': 0.485134} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.252492] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Reconfigured VM instance instance-0000004c to attach disk [datastore2] b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e/b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.253146] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af7f52ca-b4e1-4417-9991-d6a45ce45485 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.261034] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 957.261034] env[62952]: value = "task-1263494" [ 957.261034] env[62952]: _type = "Task" [ 957.261034] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.271573] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263494, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.529317] env[62952]: DEBUG oslo_vmware.api [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263485, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.544475] env[62952]: DEBUG nova.compute.manager [req-8ef4571d-a9cd-45ce-a0a4-7dc64e3ce1fb req-61571d50-e2af-4ea7-97ee-b6cca3cc8c00 service nova] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Received event network-vif-plugged-d93fe914-641a-4750-a765-181c7012b252 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 957.544475] env[62952]: DEBUG oslo_concurrency.lockutils [req-8ef4571d-a9cd-45ce-a0a4-7dc64e3ce1fb req-61571d50-e2af-4ea7-97ee-b6cca3cc8c00 service nova] Acquiring lock "dbca4369-1ed3-493a-9847-9fa1e4293475-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.544475] env[62952]: DEBUG oslo_concurrency.lockutils [req-8ef4571d-a9cd-45ce-a0a4-7dc64e3ce1fb req-61571d50-e2af-4ea7-97ee-b6cca3cc8c00 service nova] Lock "dbca4369-1ed3-493a-9847-9fa1e4293475-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.544831] env[62952]: DEBUG oslo_concurrency.lockutils [req-8ef4571d-a9cd-45ce-a0a4-7dc64e3ce1fb req-61571d50-e2af-4ea7-97ee-b6cca3cc8c00 service nova] Lock "dbca4369-1ed3-493a-9847-9fa1e4293475-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.544831] env[62952]: DEBUG nova.compute.manager [req-8ef4571d-a9cd-45ce-a0a4-7dc64e3ce1fb req-61571d50-e2af-4ea7-97ee-b6cca3cc8c00 service nova] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] No waiting events found dispatching network-vif-plugged-d93fe914-641a-4750-a765-181c7012b252 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 957.544987] env[62952]: WARNING nova.compute.manager [req-8ef4571d-a9cd-45ce-a0a4-7dc64e3ce1fb req-61571d50-e2af-4ea7-97ee-b6cca3cc8c00 service nova] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Received unexpected event network-vif-plugged-d93fe914-641a-4750-a765-181c7012b252 for instance with vm_state building and task_state spawning. [ 957.545286] env[62952]: DEBUG nova.compute.manager [req-8ef4571d-a9cd-45ce-a0a4-7dc64e3ce1fb req-61571d50-e2af-4ea7-97ee-b6cca3cc8c00 service nova] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Received event network-changed-d93fe914-641a-4750-a765-181c7012b252 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 957.545479] env[62952]: DEBUG nova.compute.manager [req-8ef4571d-a9cd-45ce-a0a4-7dc64e3ce1fb req-61571d50-e2af-4ea7-97ee-b6cca3cc8c00 service nova] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Refreshing instance network info cache due to event network-changed-d93fe914-641a-4750-a765-181c7012b252. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 957.545684] env[62952]: DEBUG oslo_concurrency.lockutils [req-8ef4571d-a9cd-45ce-a0a4-7dc64e3ce1fb req-61571d50-e2af-4ea7-97ee-b6cca3cc8c00 service nova] Acquiring lock "refresh_cache-dbca4369-1ed3-493a-9847-9fa1e4293475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.665906] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.327s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.666576] env[62952]: DEBUG nova.compute.manager [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 957.669623] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 19.607s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.669804] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.669958] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 957.670265] env[62952]: DEBUG oslo_concurrency.lockutils [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.352s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.670477] env[62952]: DEBUG nova.objects.instance [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Lazy-loading 'resources' on Instance uuid 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.672276] env[62952]: DEBUG nova.network.neutron [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 957.674595] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d2c9e0a-5bcc-4b1c-bd9f-2b5ee7b5b19e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.684564] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88891f68-5fd6-45f1-ad1e-a03d8e8c6437 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.706864] env[62952]: DEBUG oslo_vmware.api [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263492, 'name': PowerOnVM_Task, 'duration_secs': 0.595696} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.707630] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc20dc71-567b-4c48-b8da-40597173aebb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.710149] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 957.710394] env[62952]: DEBUG nova.compute.manager [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 957.711604] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b82a1a2-2952-4e21-b2f5-2e1998e1f9d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.719980] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea075f66-ad59-4d19-917c-f9c0a26f6e60 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.757694] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179950MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 957.757860] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.770955] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263494, 'name': Rename_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.876717] env[62952]: DEBUG nova.network.neutron [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Updating instance_info_cache with network_info: [{"id": "d93fe914-641a-4750-a765-181c7012b252", "address": "fa:16:3e:04:cd:36", "network": {"id": "827a8171-6759-4434-a6f1-437aad881c81", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1902486970-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091cdeb6da48492bb02f93822a45c9bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11b669be-fb26-4ef8-bdb6-c77ab9d06daf", "external-id": "nsx-vlan-transportzone-633", "segmentation_id": 633, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd93fe914-64", "ovs_interfaceid": "d93fe914-641a-4750-a765-181c7012b252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.030303] env[62952]: DEBUG oslo_vmware.api [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263485, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.174071] env[62952]: DEBUG nova.compute.utils [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 958.177282] env[62952]: DEBUG nova.compute.manager [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 958.177282] env[62952]: DEBUG nova.network.neutron [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 958.239644] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.241570] env[62952]: DEBUG nova.policy [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '23bdda22bcf74a1ea9d681512d663aeb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '091cdeb6da48492bb02f93822a45c9bf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 958.272481] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263494, 'name': Rename_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.379802] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Releasing lock "refresh_cache-dbca4369-1ed3-493a-9847-9fa1e4293475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.380141] env[62952]: DEBUG nova.compute.manager [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Instance network_info: |[{"id": "d93fe914-641a-4750-a765-181c7012b252", "address": "fa:16:3e:04:cd:36", "network": {"id": "827a8171-6759-4434-a6f1-437aad881c81", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1902486970-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091cdeb6da48492bb02f93822a45c9bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11b669be-fb26-4ef8-bdb6-c77ab9d06daf", "external-id": "nsx-vlan-transportzone-633", "segmentation_id": 633, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd93fe914-64", "ovs_interfaceid": "d93fe914-641a-4750-a765-181c7012b252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 958.380450] env[62952]: DEBUG oslo_concurrency.lockutils [req-8ef4571d-a9cd-45ce-a0a4-7dc64e3ce1fb req-61571d50-e2af-4ea7-97ee-b6cca3cc8c00 service nova] Acquired lock "refresh_cache-dbca4369-1ed3-493a-9847-9fa1e4293475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.380628] env[62952]: DEBUG nova.network.neutron [req-8ef4571d-a9cd-45ce-a0a4-7dc64e3ce1fb req-61571d50-e2af-4ea7-97ee-b6cca3cc8c00 service nova] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Refreshing network info cache for port d93fe914-641a-4750-a765-181c7012b252 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 958.381914] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:cd:36', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '11b669be-fb26-4ef8-bdb6-c77ab9d06daf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd93fe914-641a-4750-a765-181c7012b252', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 958.390717] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Creating folder: Project (091cdeb6da48492bb02f93822a45c9bf). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 958.396040] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-17160223-885a-451c-b159-bde8c211e99b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.408423] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Created folder: Project (091cdeb6da48492bb02f93822a45c9bf) in parent group-v271811. [ 958.408423] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Creating folder: Instances. Parent ref: group-v271917. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 958.409422] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ba51141-406c-44f4-8434-a70183347415 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.424072] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Created folder: Instances in parent group-v271917. [ 958.424072] env[62952]: DEBUG oslo.service.loopingcall [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.424072] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 958.424072] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ef44067-2f00-46b3-bc76-4487228bbd30 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.449206] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 958.449206] env[62952]: value = "task-1263497" [ 958.449206] env[62952]: _type = "Task" [ 958.449206] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.455902] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263497, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.490377] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d5403ea-63ec-4c39-8318-a3742a8bf370 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.502349] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4e04e1-407b-4cc2-9f48-9416f0a73d8a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.553074] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ca7b29-3ff7-4d8e-a07c-3c8d81245140 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.562997] env[62952]: DEBUG oslo_vmware.api [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263485, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.564362] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d73a80f-34fa-4463-8a86-538b2cff6182 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.579437] env[62952]: DEBUG nova.compute.provider_tree [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.637965] env[62952]: DEBUG oslo_vmware.rw_handles [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5241db49-b1ed-9adb-1eb8-ee8a1671af95/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 958.639993] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b601b7ce-4701-4230-aa1f-8902cc0a40dd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.645990] env[62952]: DEBUG oslo_vmware.rw_handles [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5241db49-b1ed-9adb-1eb8-ee8a1671af95/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 958.646228] env[62952]: ERROR oslo_vmware.rw_handles [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5241db49-b1ed-9adb-1eb8-ee8a1671af95/disk-0.vmdk due to incomplete transfer. [ 958.646488] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-bb2e4e8e-880d-4981-8770-7a59339527e4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.657149] env[62952]: DEBUG oslo_vmware.rw_handles [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5241db49-b1ed-9adb-1eb8-ee8a1671af95/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 958.657512] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Uploaded image d36ac993-3e17-46d7-a6d5-93c4255077dc to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 958.660624] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 958.660891] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c652294e-447c-4bfc-ab81-cd7b0fcd3da6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.667522] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 958.667522] env[62952]: value = "task-1263498" [ 958.667522] env[62952]: _type = "Task" [ 958.667522] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.668697] env[62952]: DEBUG nova.network.neutron [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Successfully created port: bd2fc6b2-78ee-4677-899a-db20d7a7ce8f {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 958.679520] env[62952]: DEBUG nova.compute.manager [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 958.682206] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263498, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.774459] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263494, 'name': Rename_Task, 'duration_secs': 1.270152} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.774832] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 958.775043] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1bb8e3b5-1897-4fcd-891e-8bc256cffcfc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.781436] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 958.781436] env[62952]: value = "task-1263499" [ 958.781436] env[62952]: _type = "Task" [ 958.781436] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.790659] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263499, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.846472] env[62952]: DEBUG oslo_concurrency.lockutils [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "e17b3d83-7cea-4a65-9457-7e57a6e7641a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.846777] env[62952]: DEBUG oslo_concurrency.lockutils [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "e17b3d83-7cea-4a65-9457-7e57a6e7641a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.846997] env[62952]: DEBUG oslo_concurrency.lockutils [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "e17b3d83-7cea-4a65-9457-7e57a6e7641a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.847208] env[62952]: DEBUG oslo_concurrency.lockutils [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "e17b3d83-7cea-4a65-9457-7e57a6e7641a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.847405] env[62952]: DEBUG oslo_concurrency.lockutils [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "e17b3d83-7cea-4a65-9457-7e57a6e7641a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.849893] env[62952]: INFO nova.compute.manager [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Terminating instance [ 958.852154] env[62952]: DEBUG nova.compute.manager [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 958.852552] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 958.853380] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728ca445-0c88-404a-82e6-f5e8ae36e4cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.861309] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 958.861755] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-929778a9-d148-4571-a264-8de0d14280a1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.868592] env[62952]: DEBUG oslo_vmware.api [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 958.868592] env[62952]: value = "task-1263500" [ 958.868592] env[62952]: _type = "Task" [ 958.868592] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.877745] env[62952]: DEBUG oslo_vmware.api [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263500, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.959919] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263497, 'name': CreateVM_Task, 'duration_secs': 0.49149} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.960103] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 958.960922] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.961175] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.961603] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 958.961963] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5d4001e-e2cb-4fdb-b578-2dd993455616 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.966952] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 958.966952] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52de45ba-6562-e45f-6ae7-e03b81d33911" [ 958.966952] env[62952]: _type = "Task" [ 958.966952] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.976148] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52de45ba-6562-e45f-6ae7-e03b81d33911, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.059840] env[62952]: DEBUG oslo_vmware.api [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263485, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.082653] env[62952]: DEBUG nova.scheduler.client.report [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 959.192552] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263498, 'name': Destroy_Task, 'duration_secs': 0.476319} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.195345] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Destroyed the VM [ 959.199361] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 959.199827] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e6694040-0266-407e-90ea-a7b4f9a16592 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.210636] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 959.210636] env[62952]: value = "task-1263501" [ 959.210636] env[62952]: _type = "Task" [ 959.210636] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.220237] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263501, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.294412] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263499, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.324262] env[62952]: DEBUG nova.network.neutron [req-8ef4571d-a9cd-45ce-a0a4-7dc64e3ce1fb req-61571d50-e2af-4ea7-97ee-b6cca3cc8c00 service nova] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Updated VIF entry in instance network info cache for port d93fe914-641a-4750-a765-181c7012b252. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 959.324652] env[62952]: DEBUG nova.network.neutron [req-8ef4571d-a9cd-45ce-a0a4-7dc64e3ce1fb req-61571d50-e2af-4ea7-97ee-b6cca3cc8c00 service nova] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Updating instance_info_cache with network_info: [{"id": "d93fe914-641a-4750-a765-181c7012b252", "address": "fa:16:3e:04:cd:36", "network": {"id": "827a8171-6759-4434-a6f1-437aad881c81", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1902486970-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091cdeb6da48492bb02f93822a45c9bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11b669be-fb26-4ef8-bdb6-c77ab9d06daf", "external-id": "nsx-vlan-transportzone-633", "segmentation_id": 633, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd93fe914-64", "ovs_interfaceid": "d93fe914-641a-4750-a765-181c7012b252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.381402] env[62952]: DEBUG oslo_vmware.api [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263500, 'name': PowerOffVM_Task, 'duration_secs': 0.181992} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.382290] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 959.382290] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 959.382525] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dffda28d-eea0-4013-8a44-642a82a07707 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.447800] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 959.447995] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 959.448204] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleting the datastore file [datastore2] e17b3d83-7cea-4a65-9457-7e57a6e7641a {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 959.448493] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-21315780-05ba-4164-b9a5-1baf41e8718f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.456086] env[62952]: DEBUG oslo_vmware.api [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 959.456086] env[62952]: value = "task-1263503" [ 959.456086] env[62952]: _type = "Task" [ 959.456086] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.464668] env[62952]: DEBUG oslo_vmware.api [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263503, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.479875] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52de45ba-6562-e45f-6ae7-e03b81d33911, 'name': SearchDatastore_Task, 'duration_secs': 0.011557} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.479875] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.479875] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.479875] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.479875] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.479875] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.480079] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e4a88645-fd2d-42a2-90a6-686b2463c01d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.494995] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.494995] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 959.497154] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b336f5a-0b56-4cb9-9a88-fce9f2e4effd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.501778] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 959.501778] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52320461-af1e-36d6-5a07-48fb39e27eb0" [ 959.501778] env[62952]: _type = "Task" [ 959.501778] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.509913] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52320461-af1e-36d6-5a07-48fb39e27eb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.558251] env[62952]: DEBUG oslo_vmware.api [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263485, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.588386] env[62952]: DEBUG oslo_concurrency.lockutils [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.918s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.591434] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.103s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.591598] env[62952]: DEBUG nova.objects.instance [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lazy-loading 'pci_requests' on Instance uuid 9cc7bc2f-ff22-43cf-be79-ff8c6426115b {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.609388] env[62952]: INFO nova.scheduler.client.report [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Deleted allocations for instance 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20 [ 959.694415] env[62952]: DEBUG nova.compute.manager [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 959.725161] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263501, 'name': RemoveSnapshot_Task} progress is 62%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.728866] env[62952]: DEBUG nova.virt.hardware [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 959.729137] env[62952]: DEBUG nova.virt.hardware [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 959.729302] env[62952]: DEBUG nova.virt.hardware [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 959.729487] env[62952]: DEBUG nova.virt.hardware [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 959.729640] env[62952]: DEBUG nova.virt.hardware [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 959.729789] env[62952]: DEBUG nova.virt.hardware [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 959.729998] env[62952]: DEBUG nova.virt.hardware [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 959.730372] env[62952]: DEBUG nova.virt.hardware [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 959.730573] env[62952]: DEBUG nova.virt.hardware [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 959.730742] env[62952]: DEBUG nova.virt.hardware [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 959.730915] env[62952]: DEBUG nova.virt.hardware [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.731766] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b4c494f-05f5-4415-816c-8ac5f28446fe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.744746] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61529814-254b-40ef-b2eb-5f143b18a202 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.763763] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.763881] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.791392] env[62952]: DEBUG oslo_vmware.api [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263499, 'name': PowerOnVM_Task, 'duration_secs': 0.536338} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.791679] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 959.791885] env[62952]: INFO nova.compute.manager [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Took 9.02 seconds to spawn the instance on the hypervisor. [ 959.792084] env[62952]: DEBUG nova.compute.manager [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 959.792950] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52ea168e-4e15-4435-9c3f-f4a342d082de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.830827] env[62952]: DEBUG oslo_concurrency.lockutils [req-8ef4571d-a9cd-45ce-a0a4-7dc64e3ce1fb req-61571d50-e2af-4ea7-97ee-b6cca3cc8c00 service nova] Releasing lock "refresh_cache-dbca4369-1ed3-493a-9847-9fa1e4293475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.966555] env[62952]: DEBUG oslo_vmware.api [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263503, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144606} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.966895] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.967167] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 959.967467] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 959.967716] env[62952]: INFO nova.compute.manager [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 959.968026] env[62952]: DEBUG oslo.service.loopingcall [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.968291] env[62952]: DEBUG nova.compute.manager [-] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 959.968438] env[62952]: DEBUG nova.network.neutron [-] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 960.012582] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52320461-af1e-36d6-5a07-48fb39e27eb0, 'name': SearchDatastore_Task, 'duration_secs': 0.008618} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.012582] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f5ebc9e-c4bf-4bd5-a713-24f9e1582502 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.018128] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 960.018128] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f009a2-cd2c-6f62-2080-287bb227ad2e" [ 960.018128] env[62952]: _type = "Task" [ 960.018128] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.027773] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f009a2-cd2c-6f62-2080-287bb227ad2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.059673] env[62952]: DEBUG oslo_vmware.api [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263485, 'name': ReconfigVM_Task, 'duration_secs': 5.776462} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.059933] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.060231] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Reconfigured VM to detach interface {{(pid=62952) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 960.101987] env[62952]: DEBUG nova.objects.instance [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lazy-loading 'numa_topology' on Instance uuid 9cc7bc2f-ff22-43cf-be79-ff8c6426115b {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.125434] env[62952]: DEBUG oslo_concurrency.lockutils [None req-acfe19fa-1107-4718-a2ec-262e6ea6f5dc tempest-AttachInterfacesV270Test-1066807100 tempest-AttachInterfacesV270Test-1066807100-project-member] Lock "45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.727s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.222618] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263501, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.271217] env[62952]: DEBUG nova.compute.manager [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 960.317709] env[62952]: INFO nova.compute.manager [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Took 29.94 seconds to build instance. [ 960.355289] env[62952]: DEBUG nova.compute.manager [req-7231177f-01a5-419e-a561-142dea90172b req-d3dc3c57-09a6-4289-8b74-20e555277226 service nova] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Received event network-vif-deleted-d1e678fd-723e-413a-a8e8-44e88b909665 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 960.355289] env[62952]: INFO nova.compute.manager [req-7231177f-01a5-419e-a561-142dea90172b req-d3dc3c57-09a6-4289-8b74-20e555277226 service nova] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Neutron deleted interface d1e678fd-723e-413a-a8e8-44e88b909665; detaching it from the instance and deleting it from the info cache [ 960.355289] env[62952]: DEBUG nova.network.neutron [req-7231177f-01a5-419e-a561-142dea90172b req-d3dc3c57-09a6-4289-8b74-20e555277226 service nova] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.532020] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f009a2-cd2c-6f62-2080-287bb227ad2e, 'name': SearchDatastore_Task, 'duration_secs': 0.009094} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.532020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.532020] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] dbca4369-1ed3-493a-9847-9fa1e4293475/dbca4369-1ed3-493a-9847-9fa1e4293475.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 960.532020] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9522d20-47da-457a-ae9b-3024f5df2a4a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.537691] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 960.537691] env[62952]: value = "task-1263504" [ 960.537691] env[62952]: _type = "Task" [ 960.537691] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.545767] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263504, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.609103] env[62952]: INFO nova.compute.claims [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 960.662196] env[62952]: DEBUG nova.network.neutron [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Successfully updated port: bd2fc6b2-78ee-4677-899a-db20d7a7ce8f {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 960.709414] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "7ba1587d-edaf-4dce-a224-6b2ac22cfecd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.709628] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "7ba1587d-edaf-4dce-a224-6b2ac22cfecd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.723727] env[62952]: DEBUG oslo_vmware.api [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263501, 'name': RemoveSnapshot_Task, 'duration_secs': 1.013516} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.728623] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 960.728623] env[62952]: INFO nova.compute.manager [None req-edac87cc-a0a7-4fad-a969-61bd0b681298 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Took 15.58 seconds to snapshot the instance on the hypervisor. [ 960.785126] env[62952]: DEBUG nova.network.neutron [-] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.801817] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.819612] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8f478902-de8e-497f-8df8-c8af6a35fef3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.454s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.858233] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8b3202a5-71b2-4c89-92fa-1d504eb02a54 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.873830] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d762820-f7aa-4005-95bc-18875b5af87e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.913122] env[62952]: DEBUG nova.compute.manager [req-7231177f-01a5-419e-a561-142dea90172b req-d3dc3c57-09a6-4289-8b74-20e555277226 service nova] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Detach interface failed, port_id=d1e678fd-723e-413a-a8e8-44e88b909665, reason: Instance e17b3d83-7cea-4a65-9457-7e57a6e7641a could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 961.058707] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263504, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.167628] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "refresh_cache-77fe79de-8da8-486e-af63-91f9d8196ca9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.167628] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired lock "refresh_cache-77fe79de-8da8-486e-af63-91f9d8196ca9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.167628] env[62952]: DEBUG nova.network.neutron [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 961.215225] env[62952]: DEBUG nova.compute.manager [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 961.289170] env[62952]: INFO nova.compute.manager [-] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Took 1.32 seconds to deallocate network for instance. [ 961.459663] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.459663] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.460145] env[62952]: DEBUG nova.network.neutron [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 961.551743] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263504, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.642228} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.552023] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] dbca4369-1ed3-493a-9847-9fa1e4293475/dbca4369-1ed3-493a-9847-9fa1e4293475.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 961.552376] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 961.552513] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4122a622-feb6-4752-ba42-629e2e43bf67 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.559067] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 961.559067] env[62952]: value = "task-1263505" [ 961.559067] env[62952]: _type = "Task" [ 961.559067] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.566445] env[62952]: DEBUG oslo_concurrency.lockutils [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.566922] env[62952]: DEBUG oslo_concurrency.lockutils [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.576241] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263505, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.731719] env[62952]: DEBUG nova.network.neutron [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 961.734509] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.799905] env[62952]: DEBUG oslo_concurrency.lockutils [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.907488] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be615fa4-7865-4dbb-9130-e52e3093e37a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.917377] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ed77d27-8711-4736-b905-dca880c262fa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.956832] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7407e043-0f8f-48b4-9795-ca09acfdf9e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.966797] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6d436f-70ce-4ac0-8755-569195a1f180 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.987171] env[62952]: DEBUG nova.compute.provider_tree [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.072967] env[62952]: DEBUG nova.compute.utils [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 962.074408] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263505, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068854} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.075424] env[62952]: DEBUG nova.network.neutron [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Updating instance_info_cache with network_info: [{"id": "bd2fc6b2-78ee-4677-899a-db20d7a7ce8f", "address": "fa:16:3e:89:03:5a", "network": {"id": "827a8171-6759-4434-a6f1-437aad881c81", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1902486970-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091cdeb6da48492bb02f93822a45c9bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11b669be-fb26-4ef8-bdb6-c77ab9d06daf", "external-id": "nsx-vlan-transportzone-633", "segmentation_id": 633, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd2fc6b2-78", "ovs_interfaceid": "bd2fc6b2-78ee-4677-899a-db20d7a7ce8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.077109] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 962.078157] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83fdbac-d151-4e6e-84dd-9bd87c563708 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.108377] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] dbca4369-1ed3-493a-9847-9fa1e4293475/dbca4369-1ed3-493a-9847-9fa1e4293475.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.112381] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d875db00-6189-4371-93f3-08ab9de77744 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.137669] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 962.137669] env[62952]: value = "task-1263506" [ 962.137669] env[62952]: _type = "Task" [ 962.137669] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.148494] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263506, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.335674] env[62952]: INFO nova.network.neutron [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Port bdfdd199-b7c7-47fd-bcd7-15760750128b from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 962.335674] env[62952]: DEBUG nova.network.neutron [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Updating instance_info_cache with network_info: [{"id": "b5b040c9-55a9-4929-b720-a1988f7492c9", "address": "fa:16:3e:0b:89:af", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5b040c9-55", "ovs_interfaceid": "b5b040c9-55a9-4929-b720-a1988f7492c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.495096] env[62952]: DEBUG nova.scheduler.client.report [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.514835] env[62952]: DEBUG nova.compute.manager [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Received event network-vif-plugged-bd2fc6b2-78ee-4677-899a-db20d7a7ce8f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 962.515529] env[62952]: DEBUG oslo_concurrency.lockutils [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] Acquiring lock "77fe79de-8da8-486e-af63-91f9d8196ca9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.515747] env[62952]: DEBUG oslo_concurrency.lockutils [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] Lock "77fe79de-8da8-486e-af63-91f9d8196ca9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.515948] env[62952]: DEBUG oslo_concurrency.lockutils [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] Lock "77fe79de-8da8-486e-af63-91f9d8196ca9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.516132] env[62952]: DEBUG nova.compute.manager [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] No waiting events found dispatching network-vif-plugged-bd2fc6b2-78ee-4677-899a-db20d7a7ce8f {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 962.516299] env[62952]: WARNING nova.compute.manager [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Received unexpected event network-vif-plugged-bd2fc6b2-78ee-4677-899a-db20d7a7ce8f for instance with vm_state building and task_state spawning. [ 962.516520] env[62952]: DEBUG nova.compute.manager [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Received event network-changed-bd2fc6b2-78ee-4677-899a-db20d7a7ce8f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 962.516725] env[62952]: DEBUG nova.compute.manager [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Refreshing instance network info cache due to event network-changed-bd2fc6b2-78ee-4677-899a-db20d7a7ce8f. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 962.516907] env[62952]: DEBUG oslo_concurrency.lockutils [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] Acquiring lock "refresh_cache-77fe79de-8da8-486e-af63-91f9d8196ca9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.578313] env[62952]: DEBUG oslo_concurrency.lockutils [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.579076] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Releasing lock "refresh_cache-77fe79de-8da8-486e-af63-91f9d8196ca9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.579387] env[62952]: DEBUG nova.compute.manager [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Instance network_info: |[{"id": "bd2fc6b2-78ee-4677-899a-db20d7a7ce8f", "address": "fa:16:3e:89:03:5a", "network": {"id": "827a8171-6759-4434-a6f1-437aad881c81", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1902486970-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091cdeb6da48492bb02f93822a45c9bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11b669be-fb26-4ef8-bdb6-c77ab9d06daf", "external-id": "nsx-vlan-transportzone-633", "segmentation_id": 633, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd2fc6b2-78", "ovs_interfaceid": "bd2fc6b2-78ee-4677-899a-db20d7a7ce8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 962.580954] env[62952]: DEBUG oslo_concurrency.lockutils [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] Acquired lock "refresh_cache-77fe79de-8da8-486e-af63-91f9d8196ca9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.581215] env[62952]: DEBUG nova.network.neutron [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Refreshing network info cache for port bd2fc6b2-78ee-4677-899a-db20d7a7ce8f {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 962.585308] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:03:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '11b669be-fb26-4ef8-bdb6-c77ab9d06daf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bd2fc6b2-78ee-4677-899a-db20d7a7ce8f', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 962.590677] env[62952]: DEBUG oslo.service.loopingcall [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.595525] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 962.596892] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6ea26bd6-e083-4707-bb70-6afe3624a830 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.619746] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 962.619746] env[62952]: value = "task-1263507" [ 962.619746] env[62952]: _type = "Task" [ 962.619746] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.629748] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263507, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.648570] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263506, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.839352] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.861306] env[62952]: DEBUG nova.network.neutron [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Updated VIF entry in instance network info cache for port bd2fc6b2-78ee-4677-899a-db20d7a7ce8f. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 962.861938] env[62952]: DEBUG nova.network.neutron [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Updating instance_info_cache with network_info: [{"id": "bd2fc6b2-78ee-4677-899a-db20d7a7ce8f", "address": "fa:16:3e:89:03:5a", "network": {"id": "827a8171-6759-4434-a6f1-437aad881c81", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1902486970-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091cdeb6da48492bb02f93822a45c9bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11b669be-fb26-4ef8-bdb6-c77ab9d06daf", "external-id": "nsx-vlan-transportzone-633", "segmentation_id": 633, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd2fc6b2-78", "ovs_interfaceid": "bd2fc6b2-78ee-4677-899a-db20d7a7ce8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.001200] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.410s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.003445] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.228s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.004974] env[62952]: INFO nova.compute.claims [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 963.052808] env[62952]: INFO nova.network.neutron [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Updating port 3a705583-6448-4020-9894-5f2056e2e4e7 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 963.130900] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263507, 'name': CreateVM_Task, 'duration_secs': 0.340403} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.131150] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 963.131841] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.132016] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.132355] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 963.132799] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed3cf19f-3056-427f-b5d4-65219d6d6f64 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.137106] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 963.137106] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52935eb3-e75f-7260-f92c-06da62e884a9" [ 963.137106] env[62952]: _type = "Task" [ 963.137106] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.152518] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52935eb3-e75f-7260-f92c-06da62e884a9, 'name': SearchDatastore_Task, 'duration_secs': 0.008908} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.154836] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.155084] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 963.155310] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.155455] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.155630] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 963.155902] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263506, 'name': ReconfigVM_Task, 'duration_secs': 0.651358} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.156860] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-03cd36b5-5a79-4f97-a3df-13a9546d8625 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.158841] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Reconfigured VM instance instance-0000004d to attach disk [datastore1] dbca4369-1ed3-493a-9847-9fa1e4293475/dbca4369-1ed3-493a-9847-9fa1e4293475.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.161730] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7feb6cc9-6ef7-45fd-b384-c0367f9dcd61 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.168352] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 963.168352] env[62952]: value = "task-1263508" [ 963.168352] env[62952]: _type = "Task" [ 963.168352] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.170896] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 963.171089] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 963.172270] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15972f66-fd20-4cd8-8c67-c1649f71a6ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.177997] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263508, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.180819] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 963.180819] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5266d7b3-6274-770a-0d46-770eeebb7866" [ 963.180819] env[62952]: _type = "Task" [ 963.180819] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.190851] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5266d7b3-6274-770a-0d46-770eeebb7866, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.207971] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "interface-dafa4926-3ab5-47f9-8ff2-4049e05c3113-bdfdd199-b7c7-47fd-bcd7-15760750128b" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.208523] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "interface-dafa4926-3ab5-47f9-8ff2-4049e05c3113-bdfdd199-b7c7-47fd-bcd7-15760750128b" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.208698] env[62952]: DEBUG nova.objects.instance [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lazy-loading 'flavor' on Instance uuid dafa4926-3ab5-47f9-8ff2-4049e05c3113 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.342460] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d28c2e41-f213-42e4-bb24-9f06c36bcdbe tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "interface-0533ee81-b69d-4f4e-9832-78c5ab4512b1-bdfdd199-b7c7-47fd-bcd7-15760750128b" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.924s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.365379] env[62952]: DEBUG oslo_concurrency.lockutils [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] Releasing lock "refresh_cache-77fe79de-8da8-486e-af63-91f9d8196ca9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.365716] env[62952]: DEBUG nova.compute.manager [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Received event network-changed-b5b040c9-55a9-4929-b720-a1988f7492c9 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 963.365895] env[62952]: DEBUG nova.compute.manager [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Refreshing instance network info cache due to event network-changed-b5b040c9-55a9-4929-b720-a1988f7492c9. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 963.366128] env[62952]: DEBUG oslo_concurrency.lockutils [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] Acquiring lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.366269] env[62952]: DEBUG oslo_concurrency.lockutils [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] Acquired lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.366515] env[62952]: DEBUG nova.network.neutron [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Refreshing network info cache for port b5b040c9-55a9-4929-b720-a1988f7492c9 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 963.648414] env[62952]: DEBUG oslo_concurrency.lockutils [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.648700] env[62952]: DEBUG oslo_concurrency.lockutils [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.649436] env[62952]: INFO nova.compute.manager [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Attaching volume af7338aa-c509-4319-a3a4-ce7e8aea5cfe to /dev/sdb [ 963.675690] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263508, 'name': Rename_Task, 'duration_secs': 0.136997} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.675976] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 963.676242] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6c82f0c6-a73c-4989-bfde-9ce2ebbadb4a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.680837] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b87210-d46b-4c38-97ad-bdbcd73f18ee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.684409] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 963.684409] env[62952]: value = "task-1263509" [ 963.684409] env[62952]: _type = "Task" [ 963.684409] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.692910] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad480541-ce83-467c-9521-4ceec74b9b08 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.702390] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5266d7b3-6274-770a-0d46-770eeebb7866, 'name': SearchDatastore_Task, 'duration_secs': 0.010622} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.702614] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263509, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.703868] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51868720-6522-4ac6-994f-419426e26b9c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.709044] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 963.709044] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521afc6e-80a0-82ad-1ce0-387a77b9eafd" [ 963.709044] env[62952]: _type = "Task" [ 963.709044] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.718891] env[62952]: DEBUG nova.virt.block_device [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Updating existing volume attachment record: 5e72bb52-ec06-4e42-ba06-6fc84c52c2bd {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 963.725801] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521afc6e-80a0-82ad-1ce0-387a77b9eafd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.865677] env[62952]: DEBUG nova.objects.instance [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lazy-loading 'pci_requests' on Instance uuid dafa4926-3ab5-47f9-8ff2-4049e05c3113 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.095045] env[62952]: DEBUG nova.network.neutron [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Updated VIF entry in instance network info cache for port b5b040c9-55a9-4929-b720-a1988f7492c9. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 964.095045] env[62952]: DEBUG nova.network.neutron [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Updating instance_info_cache with network_info: [{"id": "b5b040c9-55a9-4929-b720-a1988f7492c9", "address": "fa:16:3e:0b:89:af", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5b040c9-55", "ovs_interfaceid": "b5b040c9-55a9-4929-b720-a1988f7492c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.200221] env[62952]: DEBUG oslo_vmware.api [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263509, 'name': PowerOnVM_Task, 'duration_secs': 0.478593} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.200523] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 964.200691] env[62952]: INFO nova.compute.manager [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Took 8.77 seconds to spawn the instance on the hypervisor. [ 964.200875] env[62952]: DEBUG nova.compute.manager [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.201662] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb43e40f-ca46-4f7b-9bed-aeb1eb236f18 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.222542] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521afc6e-80a0-82ad-1ce0-387a77b9eafd, 'name': SearchDatastore_Task, 'duration_secs': 0.02011} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.222909] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.223202] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 77fe79de-8da8-486e-af63-91f9d8196ca9/77fe79de-8da8-486e-af63-91f9d8196ca9.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 964.223462] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b6ddd15-7a6e-45c5-8c6e-1f8cf7ff0fa4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.233366] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 964.233366] env[62952]: value = "task-1263513" [ 964.233366] env[62952]: _type = "Task" [ 964.233366] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.242853] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263513, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.281430] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e999ff44-7f25-48e3-983f-69be560b357c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.290051] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-462beae9-7cc2-4c11-aaff-22d2998635e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.321256] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323a2c20-d6d8-4330-a133-63cf45cb6ff9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.329308] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11cd4456-5b56-41c6-8b4c-25d7ef1075d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.343084] env[62952]: DEBUG nova.compute.provider_tree [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.368511] env[62952]: DEBUG nova.objects.base [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 964.368779] env[62952]: DEBUG nova.network.neutron [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 964.448534] env[62952]: DEBUG nova.policy [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6967f8ebdc944dbfb2d19b5d7acdd3e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91fde3ceb94649f698459dfc174bf835', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 964.597459] env[62952]: DEBUG oslo_concurrency.lockutils [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] Releasing lock "refresh_cache-0533ee81-b69d-4f4e-9832-78c5ab4512b1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.597690] env[62952]: DEBUG nova.compute.manager [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Received event network-changed-6bab534b-286c-4391-aeeb-fbb57655925a {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 964.597867] env[62952]: DEBUG nova.compute.manager [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Refreshing instance network info cache due to event network-changed-6bab534b-286c-4391-aeeb-fbb57655925a. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 964.598113] env[62952]: DEBUG oslo_concurrency.lockutils [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] Acquiring lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.598272] env[62952]: DEBUG oslo_concurrency.lockutils [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] Acquired lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.598451] env[62952]: DEBUG nova.network.neutron [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Refreshing network info cache for port 6bab534b-286c-4391-aeeb-fbb57655925a {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 964.643888] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.644213] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.644443] env[62952]: DEBUG nova.network.neutron [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 964.724640] env[62952]: DEBUG nova.compute.manager [req-99664652-0fbb-4cf8-a142-b785e7c71af1 req-9675bd18-b0bb-4998-97a4-5f42d3bc2d65 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Received event network-vif-plugged-3a705583-6448-4020-9894-5f2056e2e4e7 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 964.724640] env[62952]: DEBUG oslo_concurrency.lockutils [req-99664652-0fbb-4cf8-a142-b785e7c71af1 req-9675bd18-b0bb-4998-97a4-5f42d3bc2d65 service nova] Acquiring lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.724640] env[62952]: DEBUG oslo_concurrency.lockutils [req-99664652-0fbb-4cf8-a142-b785e7c71af1 req-9675bd18-b0bb-4998-97a4-5f42d3bc2d65 service nova] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.724640] env[62952]: DEBUG oslo_concurrency.lockutils [req-99664652-0fbb-4cf8-a142-b785e7c71af1 req-9675bd18-b0bb-4998-97a4-5f42d3bc2d65 service nova] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.724640] env[62952]: DEBUG nova.compute.manager [req-99664652-0fbb-4cf8-a142-b785e7c71af1 req-9675bd18-b0bb-4998-97a4-5f42d3bc2d65 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] No waiting events found dispatching network-vif-plugged-3a705583-6448-4020-9894-5f2056e2e4e7 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 964.724640] env[62952]: WARNING nova.compute.manager [req-99664652-0fbb-4cf8-a142-b785e7c71af1 req-9675bd18-b0bb-4998-97a4-5f42d3bc2d65 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Received unexpected event network-vif-plugged-3a705583-6448-4020-9894-5f2056e2e4e7 for instance with vm_state shelved_offloaded and task_state spawning. [ 964.727965] env[62952]: INFO nova.compute.manager [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Took 31.17 seconds to build instance. [ 964.747683] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263513, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.848753] env[62952]: DEBUG nova.scheduler.client.report [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 965.230429] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12ee4dc1-3dff-4b6a-9dcc-9bbc4610f493 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "dbca4369-1ed3-493a-9847-9fa1e4293475" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.676s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.243940] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263513, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.557461} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.244815] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 77fe79de-8da8-486e-af63-91f9d8196ca9/77fe79de-8da8-486e-af63-91f9d8196ca9.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 965.244815] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 965.244815] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4084bffa-e516-4f6e-9d99-48e8b7b6d658 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.251258] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 965.251258] env[62952]: value = "task-1263514" [ 965.251258] env[62952]: _type = "Task" [ 965.251258] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.262671] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263514, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.331069] env[62952]: DEBUG nova.network.neutron [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Updated VIF entry in instance network info cache for port 6bab534b-286c-4391-aeeb-fbb57655925a. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 965.331425] env[62952]: DEBUG nova.network.neutron [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Updating instance_info_cache with network_info: [{"id": "6bab534b-286c-4391-aeeb-fbb57655925a", "address": "fa:16:3e:e2:f4:40", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bab534b-28", "ovs_interfaceid": "6bab534b-286c-4391-aeeb-fbb57655925a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.353840] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.354274] env[62952]: DEBUG nova.compute.manager [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 965.356886] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.349s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.357129] env[62952]: DEBUG nova.objects.instance [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lazy-loading 'resources' on Instance uuid 86ec7b3e-34d6-45e8-87b2-c8550f035ed0 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.486952] env[62952]: DEBUG nova.network.neutron [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Updating instance_info_cache with network_info: [{"id": "3a705583-6448-4020-9894-5f2056e2e4e7", "address": "fa:16:3e:86:6e:fa", "network": {"id": "247a88e9-e8e8-46af-ac41-6777f0782fe0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1297422061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d93803f89ca64e66b8271e7b0bb48983", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a705583-64", "ovs_interfaceid": "3a705583-6448-4020-9894-5f2056e2e4e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.761950] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263514, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064723} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.761950] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 965.762313] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2356499c-ce4e-4186-9d6c-fe49ea10bd7d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.784943] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 77fe79de-8da8-486e-af63-91f9d8196ca9/77fe79de-8da8-486e-af63-91f9d8196ca9.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.785229] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73c0c923-4812-4afb-bca9-f5739ebf4076 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.804945] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 965.804945] env[62952]: value = "task-1263515" [ 965.804945] env[62952]: _type = "Task" [ 965.804945] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.813554] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263515, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.828325] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquiring lock "48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.828632] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.828996] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquiring lock "48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.829347] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.829621] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.832182] env[62952]: INFO nova.compute.manager [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Terminating instance [ 965.834092] env[62952]: DEBUG oslo_concurrency.lockutils [req-c01d96a3-900c-46f0-ab29-c9ac1b9474af req-8ce02299-1bc5-4be1-854a-1393ca61af1b service nova] Releasing lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.834849] env[62952]: DEBUG nova.compute.manager [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 965.835077] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 965.836064] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6faef399-0b07-4797-b1cf-114569e5bdf3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.843526] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 965.843753] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b90f2599-9715-4727-84f8-c324ef6e0de4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.849441] env[62952]: DEBUG oslo_vmware.api [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 965.849441] env[62952]: value = "task-1263516" [ 965.849441] env[62952]: _type = "Task" [ 965.849441] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.858207] env[62952]: DEBUG oslo_vmware.api [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263516, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.859558] env[62952]: DEBUG nova.compute.utils [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 965.864492] env[62952]: DEBUG nova.compute.manager [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 965.865032] env[62952]: DEBUG nova.network.neutron [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 965.938511] env[62952]: DEBUG nova.policy [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88572e088e184d70b2cb51633539ec49', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '26f601d669ce42dca2f875ca3a68525c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 965.991281] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Releasing lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.025504] env[62952]: DEBUG nova.virt.hardware [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='3450e826e31c01e6b5f5b26da6d9887b',container_format='bare',created_at=2024-10-10T21:35:43Z,direct_url=,disk_format='vmdk',id=315e9b97-1fa4-4f98-9bd7-cd708e6717cd,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1028608331-shelved',owner='d93803f89ca64e66b8271e7b0bb48983',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-10-10T21:35:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 966.025768] env[62952]: DEBUG nova.virt.hardware [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 966.025928] env[62952]: DEBUG nova.virt.hardware [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 966.026129] env[62952]: DEBUG nova.virt.hardware [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 966.026279] env[62952]: DEBUG nova.virt.hardware [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 966.026426] env[62952]: DEBUG nova.virt.hardware [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 966.026671] env[62952]: DEBUG nova.virt.hardware [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 966.026878] env[62952]: DEBUG nova.virt.hardware [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 966.027075] env[62952]: DEBUG nova.virt.hardware [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 966.027248] env[62952]: DEBUG nova.virt.hardware [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 966.027425] env[62952]: DEBUG nova.virt.hardware [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 966.028349] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8806dea0-422a-4438-9c1e-9ab6d3a32645 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.036171] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8653e39d-ec72-463c-a037-58f6724fb415 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.054422] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:6e:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4fcde7-8926-402a-a9b7-4878d2bc1cf6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a705583-6448-4020-9894-5f2056e2e4e7', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 966.061114] env[62952]: DEBUG oslo.service.loopingcall [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.063516] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 966.063922] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-37c0e020-5b3d-491e-8e92-1149ab9f0579 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.085114] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 966.085114] env[62952]: value = "task-1263518" [ 966.085114] env[62952]: _type = "Task" [ 966.085114] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.095769] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263518, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.138120] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7922e62e-dd7c-4632-927c-44c8e4230a04 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.146173] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ffe9fe-24cc-432d-8ebd-eb6c9e512f19 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.179672] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726405ea-2274-44fd-94ec-19329699aeb6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.187795] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7e3b7a9-aea0-4b1f-a628-2c1b929605ea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.204367] env[62952]: DEBUG nova.compute.provider_tree [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.318820] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263515, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.332956] env[62952]: DEBUG nova.compute.manager [req-e3f2bbcf-31bc-430a-95a3-a2343beea0cc req-d61f8529-68db-4ec4-837f-fdba4250d78f service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Received event network-vif-plugged-bdfdd199-b7c7-47fd-bcd7-15760750128b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 966.333193] env[62952]: DEBUG oslo_concurrency.lockutils [req-e3f2bbcf-31bc-430a-95a3-a2343beea0cc req-d61f8529-68db-4ec4-837f-fdba4250d78f service nova] Acquiring lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.333400] env[62952]: DEBUG oslo_concurrency.lockutils [req-e3f2bbcf-31bc-430a-95a3-a2343beea0cc req-d61f8529-68db-4ec4-837f-fdba4250d78f service nova] Lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.333570] env[62952]: DEBUG oslo_concurrency.lockutils [req-e3f2bbcf-31bc-430a-95a3-a2343beea0cc req-d61f8529-68db-4ec4-837f-fdba4250d78f service nova] Lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.333735] env[62952]: DEBUG nova.compute.manager [req-e3f2bbcf-31bc-430a-95a3-a2343beea0cc req-d61f8529-68db-4ec4-837f-fdba4250d78f service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] No waiting events found dispatching network-vif-plugged-bdfdd199-b7c7-47fd-bcd7-15760750128b {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 966.333901] env[62952]: WARNING nova.compute.manager [req-e3f2bbcf-31bc-430a-95a3-a2343beea0cc req-d61f8529-68db-4ec4-837f-fdba4250d78f service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Received unexpected event network-vif-plugged-bdfdd199-b7c7-47fd-bcd7-15760750128b for instance with vm_state active and task_state None. [ 966.358415] env[62952]: DEBUG oslo_vmware.api [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263516, 'name': PowerOffVM_Task, 'duration_secs': 0.364936} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.358726] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 966.358900] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 966.359179] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-153ac546-51a2-404a-a4cf-3f0b27063bdd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.361240] env[62952]: DEBUG nova.network.neutron [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Successfully updated port: bdfdd199-b7c7-47fd-bcd7-15760750128b {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 966.365420] env[62952]: DEBUG nova.compute.manager [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 966.428380] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 966.428613] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 966.428825] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Deleting the datastore file [datastore2] 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 966.429134] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e88e61d4-8a55-4420-8aee-baa352a8b546 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.436822] env[62952]: DEBUG oslo_vmware.api [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 966.436822] env[62952]: value = "task-1263520" [ 966.436822] env[62952]: _type = "Task" [ 966.436822] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.443496] env[62952]: DEBUG oslo_vmware.api [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263520, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.483644] env[62952]: DEBUG nova.network.neutron [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Successfully created port: 34b7738d-65b4-4571-bef9-bf3c5ff297ef {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 966.595943] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263518, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.708399] env[62952]: DEBUG nova.scheduler.client.report [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.788563] env[62952]: DEBUG nova.network.neutron [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Successfully created port: 6a5ee171-10c8-43aa-8b0b-12f2d7068e6e {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 966.816660] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263515, 'name': ReconfigVM_Task, 'duration_secs': 0.63631} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.816992] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 77fe79de-8da8-486e-af63-91f9d8196ca9/77fe79de-8da8-486e-af63-91f9d8196ca9.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 966.817731] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9bfa9495-4e72-4960-a6e2-5d3022f8ea05 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.823922] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 966.823922] env[62952]: value = "task-1263521" [ 966.823922] env[62952]: _type = "Task" [ 966.823922] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.832844] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263521, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.863992] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.864441] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.864734] env[62952]: DEBUG nova.network.neutron [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 966.893945] env[62952]: DEBUG nova.compute.manager [req-8de72a7f-d115-426a-96b4-970af4314c66 req-d3fa9828-64b3-4fc2-aeff-c30bc5970fc1 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Received event network-changed-3a705583-6448-4020-9894-5f2056e2e4e7 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 966.894127] env[62952]: DEBUG nova.compute.manager [req-8de72a7f-d115-426a-96b4-970af4314c66 req-d3fa9828-64b3-4fc2-aeff-c30bc5970fc1 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Refreshing instance network info cache due to event network-changed-3a705583-6448-4020-9894-5f2056e2e4e7. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 966.894346] env[62952]: DEBUG oslo_concurrency.lockutils [req-8de72a7f-d115-426a-96b4-970af4314c66 req-d3fa9828-64b3-4fc2-aeff-c30bc5970fc1 service nova] Acquiring lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.894539] env[62952]: DEBUG oslo_concurrency.lockutils [req-8de72a7f-d115-426a-96b4-970af4314c66 req-d3fa9828-64b3-4fc2-aeff-c30bc5970fc1 service nova] Acquired lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.894830] env[62952]: DEBUG nova.network.neutron [req-8de72a7f-d115-426a-96b4-970af4314c66 req-d3fa9828-64b3-4fc2-aeff-c30bc5970fc1 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Refreshing network info cache for port 3a705583-6448-4020-9894-5f2056e2e4e7 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 966.945382] env[62952]: DEBUG oslo_vmware.api [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263520, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.321474} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.945650] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 966.945881] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 966.946051] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 966.946209] env[62952]: INFO nova.compute.manager [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Took 1.11 seconds to destroy the instance on the hypervisor. [ 966.946443] env[62952]: DEBUG oslo.service.loopingcall [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.946590] env[62952]: DEBUG nova.compute.manager [-] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 966.946820] env[62952]: DEBUG nova.network.neutron [-] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 967.096800] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263518, 'name': CreateVM_Task, 'duration_secs': 0.527567} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.096981] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 967.097720] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/315e9b97-1fa4-4f98-9bd7-cd708e6717cd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.098543] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired lock "[datastore1] devstack-image-cache_base/315e9b97-1fa4-4f98-9bd7-cd708e6717cd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.098543] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/315e9b97-1fa4-4f98-9bd7-cd708e6717cd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 967.098543] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e0768c5-03cf-413a-abe8-ac58e2ee799a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.103140] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 967.103140] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523b6cc7-5d0c-45dc-1270-53125c17b028" [ 967.103140] env[62952]: _type = "Task" [ 967.103140] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.112779] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523b6cc7-5d0c-45dc-1270-53125c17b028, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.214181] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.857s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.217432] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.237s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.219038] env[62952]: INFO nova.compute.claims [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 967.239963] env[62952]: INFO nova.scheduler.client.report [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Deleted allocations for instance 86ec7b3e-34d6-45e8-87b2-c8550f035ed0 [ 967.335495] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263521, 'name': Rename_Task, 'duration_secs': 0.179313} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.335799] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 967.336070] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c13a79e-b995-4838-ac62-4d8a6a1e0a1c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.342717] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 967.342717] env[62952]: value = "task-1263522" [ 967.342717] env[62952]: _type = "Task" [ 967.342717] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.351455] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263522, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.373706] env[62952]: DEBUG nova.compute.manager [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 967.400281] env[62952]: DEBUG nova.virt.hardware [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 967.400563] env[62952]: DEBUG nova.virt.hardware [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 967.400722] env[62952]: DEBUG nova.virt.hardware [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 967.401104] env[62952]: DEBUG nova.virt.hardware [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 967.401293] env[62952]: DEBUG nova.virt.hardware [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 967.401448] env[62952]: DEBUG nova.virt.hardware [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 967.401666] env[62952]: DEBUG nova.virt.hardware [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 967.401856] env[62952]: DEBUG nova.virt.hardware [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 967.401995] env[62952]: DEBUG nova.virt.hardware [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 967.402172] env[62952]: DEBUG nova.virt.hardware [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 967.402347] env[62952]: DEBUG nova.virt.hardware [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 967.405352] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ae2247-c06a-4308-bee5-03931b12a6ee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.408739] env[62952]: WARNING nova.network.neutron [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] 52af4564-da88-4cc8-b51b-07855da20a7c already exists in list: networks containing: ['52af4564-da88-4cc8-b51b-07855da20a7c']. ignoring it [ 967.415447] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53bc9298-7a55-4948-a10f-516a688f11f5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.482444] env[62952]: DEBUG nova.network.neutron [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Successfully created port: b0b3ad2c-5b56-4239-856d-663404db8324 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 967.616123] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Releasing lock "[datastore1] devstack-image-cache_base/315e9b97-1fa4-4f98-9bd7-cd708e6717cd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.616123] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Processing image 315e9b97-1fa4-4f98-9bd7-cd708e6717cd {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 967.616123] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/315e9b97-1fa4-4f98-9bd7-cd708e6717cd/315e9b97-1fa4-4f98-9bd7-cd708e6717cd.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.616123] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired lock "[datastore1] devstack-image-cache_base/315e9b97-1fa4-4f98-9bd7-cd708e6717cd/315e9b97-1fa4-4f98-9bd7-cd708e6717cd.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.616123] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 967.616315] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-56aed322-b9bb-444b-b3ac-9099252b4c03 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.628194] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 967.628398] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 967.629234] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-570624d7-5d27-4104-871a-82ebe822185c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.634916] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 967.634916] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52817082-f326-e731-bbd4-1772249ca002" [ 967.634916] env[62952]: _type = "Task" [ 967.634916] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.642138] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52817082-f326-e731-bbd4-1772249ca002, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.748878] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f809740-5318-4200-b4d0-ec83099650d1 tempest-ServerShowV247Test-884985382 tempest-ServerShowV247Test-884985382-project-member] Lock "86ec7b3e-34d6-45e8-87b2-c8550f035ed0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.004s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.781712] env[62952]: DEBUG nova.network.neutron [req-8de72a7f-d115-426a-96b4-970af4314c66 req-d3fa9828-64b3-4fc2-aeff-c30bc5970fc1 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Updated VIF entry in instance network info cache for port 3a705583-6448-4020-9894-5f2056e2e4e7. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 967.782102] env[62952]: DEBUG nova.network.neutron [req-8de72a7f-d115-426a-96b4-970af4314c66 req-d3fa9828-64b3-4fc2-aeff-c30bc5970fc1 service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Updating instance_info_cache with network_info: [{"id": "3a705583-6448-4020-9894-5f2056e2e4e7", "address": "fa:16:3e:86:6e:fa", "network": {"id": "247a88e9-e8e8-46af-ac41-6777f0782fe0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1297422061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d93803f89ca64e66b8271e7b0bb48983", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a705583-64", "ovs_interfaceid": "3a705583-6448-4020-9894-5f2056e2e4e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.853434] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263522, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.863718] env[62952]: DEBUG nova.network.neutron [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Updating instance_info_cache with network_info: [{"id": "6bab534b-286c-4391-aeeb-fbb57655925a", "address": "fa:16:3e:e2:f4:40", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bab534b-28", "ovs_interfaceid": "6bab534b-286c-4391-aeeb-fbb57655925a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bdfdd199-b7c7-47fd-bcd7-15760750128b", "address": "fa:16:3e:08:b0:ff", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdfdd199-b7", "ovs_interfaceid": "bdfdd199-b7c7-47fd-bcd7-15760750128b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.146103] env[62952]: DEBUG nova.network.neutron [-] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.147412] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Preparing fetch location {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 968.148043] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Fetch image to [datastore1] OSTACK_IMG_afa09349-4793-4e43-8ef6-1ebac150fc2a/OSTACK_IMG_afa09349-4793-4e43-8ef6-1ebac150fc2a.vmdk {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 968.148043] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Downloading stream optimized image 315e9b97-1fa4-4f98-9bd7-cd708e6717cd to [datastore1] OSTACK_IMG_afa09349-4793-4e43-8ef6-1ebac150fc2a/OSTACK_IMG_afa09349-4793-4e43-8ef6-1ebac150fc2a.vmdk on the data store datastore1 as vApp {{(pid=62952) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 968.148043] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Downloading image file data 315e9b97-1fa4-4f98-9bd7-cd708e6717cd to the ESX as VM named 'OSTACK_IMG_afa09349-4793-4e43-8ef6-1ebac150fc2a' {{(pid=62952) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 968.227644] env[62952]: DEBUG oslo_vmware.rw_handles [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 968.227644] env[62952]: value = "resgroup-9" [ 968.227644] env[62952]: _type = "ResourcePool" [ 968.227644] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 968.230715] env[62952]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-14896cd7-23e8-4914-8319-70092c4a5929 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.252187] env[62952]: DEBUG oslo_vmware.rw_handles [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lease: (returnval){ [ 968.252187] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ea80a2-79f6-de54-c26d-ddc82873ec0d" [ 968.252187] env[62952]: _type = "HttpNfcLease" [ 968.252187] env[62952]: } obtained for vApp import into resource pool (val){ [ 968.252187] env[62952]: value = "resgroup-9" [ 968.252187] env[62952]: _type = "ResourcePool" [ 968.252187] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 968.252503] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the lease: (returnval){ [ 968.252503] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ea80a2-79f6-de54-c26d-ddc82873ec0d" [ 968.252503] env[62952]: _type = "HttpNfcLease" [ 968.252503] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 968.258631] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 968.258631] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ea80a2-79f6-de54-c26d-ddc82873ec0d" [ 968.258631] env[62952]: _type = "HttpNfcLease" [ 968.258631] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 968.269859] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 968.271292] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271922', 'volume_id': 'af7338aa-c509-4319-a3a4-ce7e8aea5cfe', 'name': 'volume-af7338aa-c509-4319-a3a4-ce7e8aea5cfe', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e', 'attached_at': '', 'detached_at': '', 'volume_id': 'af7338aa-c509-4319-a3a4-ce7e8aea5cfe', 'serial': 'af7338aa-c509-4319-a3a4-ce7e8aea5cfe'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 968.272235] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40edab14-e94a-4083-aa4a-8ba039bdf862 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.288573] env[62952]: DEBUG oslo_concurrency.lockutils [req-8de72a7f-d115-426a-96b4-970af4314c66 req-d3fa9828-64b3-4fc2-aeff-c30bc5970fc1 service nova] Releasing lock "refresh_cache-9cc7bc2f-ff22-43cf-be79-ff8c6426115b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.292190] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e13135bb-8bd1-4196-ba30-919b47b8a257 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.318338] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] volume-af7338aa-c509-4319-a3a4-ce7e8aea5cfe/volume-af7338aa-c509-4319-a3a4-ce7e8aea5cfe.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.321396] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c19d354-aeb4-44cc-b861-8b8a345c2db3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.340811] env[62952]: DEBUG oslo_vmware.api [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 968.340811] env[62952]: value = "task-1263524" [ 968.340811] env[62952]: _type = "Task" [ 968.340811] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.353841] env[62952]: DEBUG oslo_vmware.api [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263524, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.358449] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263522, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.367042] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.367779] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.367940] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.370230] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-626d4f89-e1f7-49fb-bca9-c162f4f5baf1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.374856] env[62952]: DEBUG nova.compute.manager [req-796e1e42-ae7c-4ef7-99db-bfd21ee28c6f req-8220d67d-022c-4bc3-a996-658857b70614 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Received event network-changed-bdfdd199-b7c7-47fd-bcd7-15760750128b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 968.374856] env[62952]: DEBUG nova.compute.manager [req-796e1e42-ae7c-4ef7-99db-bfd21ee28c6f req-8220d67d-022c-4bc3-a996-658857b70614 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Refreshing instance network info cache due to event network-changed-bdfdd199-b7c7-47fd-bcd7-15760750128b. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 968.375181] env[62952]: DEBUG oslo_concurrency.lockutils [req-796e1e42-ae7c-4ef7-99db-bfd21ee28c6f req-8220d67d-022c-4bc3-a996-658857b70614 service nova] Acquiring lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.375213] env[62952]: DEBUG oslo_concurrency.lockutils [req-796e1e42-ae7c-4ef7-99db-bfd21ee28c6f req-8220d67d-022c-4bc3-a996-658857b70614 service nova] Acquired lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.375389] env[62952]: DEBUG nova.network.neutron [req-796e1e42-ae7c-4ef7-99db-bfd21ee28c6f req-8220d67d-022c-4bc3-a996-658857b70614 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Refreshing network info cache for port bdfdd199-b7c7-47fd-bcd7-15760750128b {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 968.392573] env[62952]: DEBUG nova.virt.hardware [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 968.392878] env[62952]: DEBUG nova.virt.hardware [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 968.393064] env[62952]: DEBUG nova.virt.hardware [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 968.393268] env[62952]: DEBUG nova.virt.hardware [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 968.393433] env[62952]: DEBUG nova.virt.hardware [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 968.393587] env[62952]: DEBUG nova.virt.hardware [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 968.393818] env[62952]: DEBUG nova.virt.hardware [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 968.394015] env[62952]: DEBUG nova.virt.hardware [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 968.394224] env[62952]: DEBUG nova.virt.hardware [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 968.394426] env[62952]: DEBUG nova.virt.hardware [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 968.394644] env[62952]: DEBUG nova.virt.hardware [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 968.401995] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Reconfiguring VM to attach interface {{(pid=62952) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 968.406655] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25dd819e-2e98-4112-bf87-7eebe30677c8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.425793] env[62952]: DEBUG oslo_vmware.api [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 968.425793] env[62952]: value = "task-1263525" [ 968.425793] env[62952]: _type = "Task" [ 968.425793] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.441549] env[62952]: DEBUG oslo_vmware.api [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263525, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.591464] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4266c051-9d14-477d-a2de-27d2e2638e42 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.600897] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8125ba-5465-487a-aecd-c68da9da84c4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.633355] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8258dec-798c-440c-bf31-3a1bd3bdd6bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.641564] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-164ddb74-9a91-48e8-bb5f-16305b7f290c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.656064] env[62952]: INFO nova.compute.manager [-] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Took 1.71 seconds to deallocate network for instance. [ 968.656643] env[62952]: DEBUG nova.compute.provider_tree [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.761719] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 968.761719] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ea80a2-79f6-de54-c26d-ddc82873ec0d" [ 968.761719] env[62952]: _type = "HttpNfcLease" [ 968.761719] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 968.762437] env[62952]: DEBUG oslo_vmware.rw_handles [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 968.762437] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ea80a2-79f6-de54-c26d-ddc82873ec0d" [ 968.762437] env[62952]: _type = "HttpNfcLease" [ 968.762437] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 968.762823] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-629fe0ba-be32-49ec-8260-fddbcd71201a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.771283] env[62952]: DEBUG oslo_vmware.rw_handles [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b1ee78-e219-3dd4-50d5-a6b209c41cac/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 968.771503] env[62952]: DEBUG oslo_vmware.rw_handles [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b1ee78-e219-3dd4-50d5-a6b209c41cac/disk-0.vmdk. {{(pid=62952) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 968.859197] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2e3f020e-373c-4a77-a1fe-98d33212d56f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.860962] env[62952]: DEBUG oslo_vmware.api [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263524, 'name': ReconfigVM_Task, 'duration_secs': 0.456183} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.866067] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Reconfigured VM instance instance-0000004c to attach disk [datastore1] volume-af7338aa-c509-4319-a3a4-ce7e8aea5cfe/volume-af7338aa-c509-4319-a3a4-ce7e8aea5cfe.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 968.870825] env[62952]: DEBUG oslo_vmware.api [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263522, 'name': PowerOnVM_Task, 'duration_secs': 1.284506} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.871439] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e7defe7-7f4c-4dd6-9853-021d06f97489 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.881543] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 968.881731] env[62952]: INFO nova.compute.manager [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Took 9.19 seconds to spawn the instance on the hypervisor. [ 968.881926] env[62952]: DEBUG nova.compute.manager [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 968.885761] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d84cfdde-350d-466e-b3e8-c97dd3721e20 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.897117] env[62952]: DEBUG oslo_vmware.api [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 968.897117] env[62952]: value = "task-1263526" [ 968.897117] env[62952]: _type = "Task" [ 968.897117] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.906459] env[62952]: DEBUG oslo_vmware.api [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263526, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.940552] env[62952]: DEBUG oslo_vmware.api [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263525, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.132413] env[62952]: DEBUG nova.network.neutron [req-796e1e42-ae7c-4ef7-99db-bfd21ee28c6f req-8220d67d-022c-4bc3-a996-658857b70614 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Updated VIF entry in instance network info cache for port bdfdd199-b7c7-47fd-bcd7-15760750128b. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 969.132987] env[62952]: DEBUG nova.network.neutron [req-796e1e42-ae7c-4ef7-99db-bfd21ee28c6f req-8220d67d-022c-4bc3-a996-658857b70614 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Updating instance_info_cache with network_info: [{"id": "6bab534b-286c-4391-aeeb-fbb57655925a", "address": "fa:16:3e:e2:f4:40", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bab534b-28", "ovs_interfaceid": "6bab534b-286c-4391-aeeb-fbb57655925a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bdfdd199-b7c7-47fd-bcd7-15760750128b", "address": "fa:16:3e:08:b0:ff", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdfdd199-b7", "ovs_interfaceid": "bdfdd199-b7c7-47fd-bcd7-15760750128b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.160700] env[62952]: DEBUG nova.scheduler.client.report [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.166556] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.279227] env[62952]: DEBUG nova.network.neutron [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Successfully updated port: 34b7738d-65b4-4571-bef9-bf3c5ff297ef {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 969.412349] env[62952]: INFO nova.compute.manager [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Took 33.74 seconds to build instance. [ 969.417837] env[62952]: DEBUG oslo_vmware.api [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263526, 'name': ReconfigVM_Task, 'duration_secs': 0.174736} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.419651] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271922', 'volume_id': 'af7338aa-c509-4319-a3a4-ce7e8aea5cfe', 'name': 'volume-af7338aa-c509-4319-a3a4-ce7e8aea5cfe', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e', 'attached_at': '', 'detached_at': '', 'volume_id': 'af7338aa-c509-4319-a3a4-ce7e8aea5cfe', 'serial': 'af7338aa-c509-4319-a3a4-ce7e8aea5cfe'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 969.437353] env[62952]: DEBUG oslo_vmware.api [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263525, 'name': ReconfigVM_Task, 'duration_secs': 0.939059} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.438915] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.439162] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Reconfigured VM to attach interface {{(pid=62952) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 969.635954] env[62952]: DEBUG oslo_concurrency.lockutils [req-796e1e42-ae7c-4ef7-99db-bfd21ee28c6f req-8220d67d-022c-4bc3-a996-658857b70614 service nova] Releasing lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.636221] env[62952]: DEBUG nova.compute.manager [req-796e1e42-ae7c-4ef7-99db-bfd21ee28c6f req-8220d67d-022c-4bc3-a996-658857b70614 service nova] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Received event network-vif-deleted-31694ebf-fc67-4662-98f6-9d99b17eb865 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 969.668523] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.669228] env[62952]: DEBUG nova.compute.manager [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 969.674886] env[62952]: DEBUG oslo_concurrency.lockutils [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.410s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.675165] env[62952]: DEBUG nova.objects.instance [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lazy-loading 'resources' on Instance uuid 5f1eca6f-4cc4-483e-b72a-b479378277fe {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.897591] env[62952]: INFO nova.compute.manager [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Rescuing [ 969.897845] env[62952]: DEBUG oslo_concurrency.lockutils [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "refresh_cache-77fe79de-8da8-486e-af63-91f9d8196ca9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.898029] env[62952]: DEBUG oslo_concurrency.lockutils [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired lock "refresh_cache-77fe79de-8da8-486e-af63-91f9d8196ca9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.898220] env[62952]: DEBUG nova.network.neutron [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 969.914357] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43a81475-727b-47fd-b829-b99ea49631fd tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "77fe79de-8da8-486e-af63-91f9d8196ca9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.254s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.944039] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50bb15e7-4339-4aa7-b69e-1234a02a380f tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "interface-dafa4926-3ab5-47f9-8ff2-4049e05c3113-bdfdd199-b7c7-47fd-bcd7-15760750128b" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.736s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.099411] env[62952]: DEBUG oslo_vmware.rw_handles [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Completed reading data from the image iterator. {{(pid=62952) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 970.099559] env[62952]: DEBUG oslo_vmware.rw_handles [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b1ee78-e219-3dd4-50d5-a6b209c41cac/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 970.100458] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e7a9919-d9b5-4b6b-9988-2dee5dbab81c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.107236] env[62952]: DEBUG oslo_vmware.rw_handles [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b1ee78-e219-3dd4-50d5-a6b209c41cac/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 970.107523] env[62952]: DEBUG oslo_vmware.rw_handles [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b1ee78-e219-3dd4-50d5-a6b209c41cac/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 970.107781] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-7b2aa4bc-50a4-400e-b84e-f5b4e258c7ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.178498] env[62952]: DEBUG nova.compute.utils [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 970.185018] env[62952]: DEBUG nova.compute.manager [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 970.185227] env[62952]: DEBUG nova.network.neutron [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 970.229984] env[62952]: DEBUG nova.policy [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9f6f0133f4a4ac287d4d8d3f4992754', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1487c06883e444a589d9b688155dc00e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 970.287114] env[62952]: DEBUG oslo_vmware.rw_handles [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b1ee78-e219-3dd4-50d5-a6b209c41cac/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 970.287365] env[62952]: INFO nova.virt.vmwareapi.images [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Downloaded image file data 315e9b97-1fa4-4f98-9bd7-cd708e6717cd [ 970.288346] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c48502-38a1-4553-9c61-ddad114211bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.307582] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-47a1c34d-6df9-44c9-8145-cdd9be859592 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.338741] env[62952]: INFO nova.virt.vmwareapi.images [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] The imported VM was unregistered [ 970.341087] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Caching image {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 970.341331] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Creating directory with path [datastore1] devstack-image-cache_base/315e9b97-1fa4-4f98-9bd7-cd708e6717cd {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 970.341707] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e307ff5f-5943-4c4a-b872-ecbd97c77cfd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.374020] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Created directory with path [datastore1] devstack-image-cache_base/315e9b97-1fa4-4f98-9bd7-cd708e6717cd {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 970.374283] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_afa09349-4793-4e43-8ef6-1ebac150fc2a/OSTACK_IMG_afa09349-4793-4e43-8ef6-1ebac150fc2a.vmdk to [datastore1] devstack-image-cache_base/315e9b97-1fa4-4f98-9bd7-cd708e6717cd/315e9b97-1fa4-4f98-9bd7-cd708e6717cd.vmdk. {{(pid=62952) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 970.374607] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-9324b311-344a-4f71-a1ce-334d740cc815 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.386020] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 970.386020] env[62952]: value = "task-1263528" [ 970.386020] env[62952]: _type = "Task" [ 970.386020] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.399224] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263528, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.404129] env[62952]: DEBUG nova.compute.manager [req-0c3b1e0c-5e64-401b-bf5d-ae12582742cf req-2b281ded-2b38-4b6c-a05d-52bd7439483c service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Received event network-vif-plugged-34b7738d-65b4-4571-bef9-bf3c5ff297ef {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 970.404336] env[62952]: DEBUG oslo_concurrency.lockutils [req-0c3b1e0c-5e64-401b-bf5d-ae12582742cf req-2b281ded-2b38-4b6c-a05d-52bd7439483c service nova] Acquiring lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.404828] env[62952]: DEBUG oslo_concurrency.lockutils [req-0c3b1e0c-5e64-401b-bf5d-ae12582742cf req-2b281ded-2b38-4b6c-a05d-52bd7439483c service nova] Lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.405044] env[62952]: DEBUG oslo_concurrency.lockutils [req-0c3b1e0c-5e64-401b-bf5d-ae12582742cf req-2b281ded-2b38-4b6c-a05d-52bd7439483c service nova] Lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.405225] env[62952]: DEBUG nova.compute.manager [req-0c3b1e0c-5e64-401b-bf5d-ae12582742cf req-2b281ded-2b38-4b6c-a05d-52bd7439483c service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] No waiting events found dispatching network-vif-plugged-34b7738d-65b4-4571-bef9-bf3c5ff297ef {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 970.405400] env[62952]: WARNING nova.compute.manager [req-0c3b1e0c-5e64-401b-bf5d-ae12582742cf req-2b281ded-2b38-4b6c-a05d-52bd7439483c service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Received unexpected event network-vif-plugged-34b7738d-65b4-4571-bef9-bf3c5ff297ef for instance with vm_state building and task_state spawning. [ 970.405563] env[62952]: DEBUG nova.compute.manager [req-0c3b1e0c-5e64-401b-bf5d-ae12582742cf req-2b281ded-2b38-4b6c-a05d-52bd7439483c service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Received event network-changed-34b7738d-65b4-4571-bef9-bf3c5ff297ef {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 970.405717] env[62952]: DEBUG nova.compute.manager [req-0c3b1e0c-5e64-401b-bf5d-ae12582742cf req-2b281ded-2b38-4b6c-a05d-52bd7439483c service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Refreshing instance network info cache due to event network-changed-34b7738d-65b4-4571-bef9-bf3c5ff297ef. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 970.405904] env[62952]: DEBUG oslo_concurrency.lockutils [req-0c3b1e0c-5e64-401b-bf5d-ae12582742cf req-2b281ded-2b38-4b6c-a05d-52bd7439483c service nova] Acquiring lock "refresh_cache-fba2e955-8b5d-461c-997b-ef8f3ec0d906" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.406053] env[62952]: DEBUG oslo_concurrency.lockutils [req-0c3b1e0c-5e64-401b-bf5d-ae12582742cf req-2b281ded-2b38-4b6c-a05d-52bd7439483c service nova] Acquired lock "refresh_cache-fba2e955-8b5d-461c-997b-ef8f3ec0d906" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.406216] env[62952]: DEBUG nova.network.neutron [req-0c3b1e0c-5e64-401b-bf5d-ae12582742cf req-2b281ded-2b38-4b6c-a05d-52bd7439483c service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Refreshing network info cache for port 34b7738d-65b4-4571-bef9-bf3c5ff297ef {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 970.422331] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbfde371-eae4-4dae-9461-49847868b376 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.430218] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c429dc2-5248-4b8a-aa60-98ce7235b4de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.465460] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f3591c3-4007-4c0f-8f0f-89091d8d46c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.468488] env[62952]: DEBUG nova.objects.instance [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lazy-loading 'flavor' on Instance uuid b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.475027] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fb673d3-bd21-42cd-b168-8384363e6382 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.490245] env[62952]: DEBUG nova.compute.provider_tree [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.538313] env[62952]: DEBUG nova.network.neutron [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Successfully created port: 89ebaa6b-2763-41e0-a0d7-954adf9ac031 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 970.687922] env[62952]: DEBUG nova.compute.manager [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 970.773263] env[62952]: DEBUG nova.network.neutron [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Updating instance_info_cache with network_info: [{"id": "bd2fc6b2-78ee-4677-899a-db20d7a7ce8f", "address": "fa:16:3e:89:03:5a", "network": {"id": "827a8171-6759-4434-a6f1-437aad881c81", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1902486970-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091cdeb6da48492bb02f93822a45c9bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11b669be-fb26-4ef8-bdb6-c77ab9d06daf", "external-id": "nsx-vlan-transportzone-633", "segmentation_id": 633, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd2fc6b2-78", "ovs_interfaceid": "bd2fc6b2-78ee-4677-899a-db20d7a7ce8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.894338] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263528, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.944540] env[62952]: DEBUG nova.network.neutron [req-0c3b1e0c-5e64-401b-bf5d-ae12582742cf req-2b281ded-2b38-4b6c-a05d-52bd7439483c service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 970.973484] env[62952]: DEBUG oslo_concurrency.lockutils [None req-568be35a-7639-4a7d-8dd9-802d5739da1c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.325s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.993115] env[62952]: DEBUG nova.scheduler.client.report [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.041385] env[62952]: DEBUG nova.network.neutron [req-0c3b1e0c-5e64-401b-bf5d-ae12582742cf req-2b281ded-2b38-4b6c-a05d-52bd7439483c service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.078807] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.078914] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.079371] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.079655] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.079849] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.082667] env[62952]: INFO nova.compute.manager [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Terminating instance [ 971.085016] env[62952]: DEBUG nova.compute.manager [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 971.085270] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 971.085542] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-980c5423-7a8e-44f1-a923-c54148160c59 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.093181] env[62952]: DEBUG oslo_vmware.api [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 971.093181] env[62952]: value = "task-1263529" [ 971.093181] env[62952]: _type = "Task" [ 971.093181] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.102420] env[62952]: DEBUG oslo_vmware.api [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263529, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.275740] env[62952]: DEBUG oslo_concurrency.lockutils [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Releasing lock "refresh_cache-77fe79de-8da8-486e-af63-91f9d8196ca9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.397336] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263528, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.418744] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fcc259d7-b371-42aa-bfb0-c0446f0ea279 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.418744] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fcc259d7-b371-42aa-bfb0-c0446f0ea279 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.418965] env[62952]: INFO nova.compute.manager [None req-fcc259d7-b371-42aa-bfb0-c0446f0ea279 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Rebooting instance [ 971.501937] env[62952]: DEBUG oslo_concurrency.lockutils [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.504782] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.623s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.505232] env[62952]: DEBUG nova.objects.instance [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lazy-loading 'resources' on Instance uuid 20ea323b-8961-4981-8045-f9cb85815d99 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.535274] env[62952]: INFO nova.scheduler.client.report [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Deleted allocations for instance 5f1eca6f-4cc4-483e-b72a-b479378277fe [ 971.544510] env[62952]: DEBUG oslo_concurrency.lockutils [req-0c3b1e0c-5e64-401b-bf5d-ae12582742cf req-2b281ded-2b38-4b6c-a05d-52bd7439483c service nova] Releasing lock "refresh_cache-fba2e955-8b5d-461c-997b-ef8f3ec0d906" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.603381] env[62952]: DEBUG oslo_vmware.api [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263529, 'name': PowerOffVM_Task, 'duration_secs': 0.27334} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.603750] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 971.603963] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 971.604223] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271922', 'volume_id': 'af7338aa-c509-4319-a3a4-ce7e8aea5cfe', 'name': 'volume-af7338aa-c509-4319-a3a4-ce7e8aea5cfe', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e', 'attached_at': '', 'detached_at': '', 'volume_id': 'af7338aa-c509-4319-a3a4-ce7e8aea5cfe', 'serial': 'af7338aa-c509-4319-a3a4-ce7e8aea5cfe'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 971.605171] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cacc24f-0049-4da4-b56e-cef07aea7036 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.628320] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a76a2c90-db7c-4d93-819f-ffc0a6956ab6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.636323] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf05d27-da8c-496d-887d-ff43305b17ba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.657829] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a47ceea-440f-42c2-bdf4-9b3a5546ec1e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.674847] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] The volume has not been displaced from its original location: [datastore1] volume-af7338aa-c509-4319-a3a4-ce7e8aea5cfe/volume-af7338aa-c509-4319-a3a4-ce7e8aea5cfe.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 971.680518] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Reconfiguring VM instance instance-0000004c to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 971.681664] env[62952]: DEBUG nova.network.neutron [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Successfully updated port: 6a5ee171-10c8-43aa-8b0b-12f2d7068e6e {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 971.682857] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b9bc7a4-2dda-4d83-b567-a7a4dc564e3b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.700688] env[62952]: DEBUG nova.compute.manager [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 971.704595] env[62952]: DEBUG oslo_vmware.api [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 971.704595] env[62952]: value = "task-1263530" [ 971.704595] env[62952]: _type = "Task" [ 971.704595] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.715605] env[62952]: DEBUG oslo_vmware.api [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263530, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.727856] env[62952]: DEBUG nova.virt.hardware [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 971.728183] env[62952]: DEBUG nova.virt.hardware [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 971.728405] env[62952]: DEBUG nova.virt.hardware [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 971.728582] env[62952]: DEBUG nova.virt.hardware [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 971.728680] env[62952]: DEBUG nova.virt.hardware [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 971.728829] env[62952]: DEBUG nova.virt.hardware [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 971.729093] env[62952]: DEBUG nova.virt.hardware [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 971.729224] env[62952]: DEBUG nova.virt.hardware [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 971.729395] env[62952]: DEBUG nova.virt.hardware [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 971.729570] env[62952]: DEBUG nova.virt.hardware [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 971.729737] env[62952]: DEBUG nova.virt.hardware [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 971.730667] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b923849c-4f80-4df0-bfb5-4c0beec8589c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.739316] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18243900-6755-4001-bf12-06220bf792e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.794989] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "interface-dafa4926-3ab5-47f9-8ff2-4049e05c3113-bdfdd199-b7c7-47fd-bcd7-15760750128b" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.795332] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "interface-dafa4926-3ab5-47f9-8ff2-4049e05c3113-bdfdd199-b7c7-47fd-bcd7-15760750128b" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.817207] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 971.817486] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7f2ee73c-8f66-4a2e-b24d-d7f9a8ad63da {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.824867] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 971.824867] env[62952]: value = "task-1263531" [ 971.824867] env[62952]: _type = "Task" [ 971.824867] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.834207] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263531, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.896679] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263528, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.945968] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fcc259d7-b371-42aa-bfb0-c0446f0ea279 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.946203] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fcc259d7-b371-42aa-bfb0-c0446f0ea279 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.946470] env[62952]: DEBUG nova.network.neutron [None req-fcc259d7-b371-42aa-bfb0-c0446f0ea279 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 972.045260] env[62952]: DEBUG oslo_concurrency.lockutils [None req-557078a4-b6f7-46fa-a8de-326d5a89cb32 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "5f1eca6f-4cc4-483e-b72a-b479378277fe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.340s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.084230] env[62952]: DEBUG nova.compute.manager [req-1f01050d-3316-4f41-95fe-92cb359b8988 req-a4c79c27-31d7-4de9-9db0-271be032a5d7 service nova] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Received event network-vif-plugged-89ebaa6b-2763-41e0-a0d7-954adf9ac031 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 972.084473] env[62952]: DEBUG oslo_concurrency.lockutils [req-1f01050d-3316-4f41-95fe-92cb359b8988 req-a4c79c27-31d7-4de9-9db0-271be032a5d7 service nova] Acquiring lock "c7d08234-fcef-4cdb-939f-05ea9e8b08ea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.084830] env[62952]: DEBUG oslo_concurrency.lockutils [req-1f01050d-3316-4f41-95fe-92cb359b8988 req-a4c79c27-31d7-4de9-9db0-271be032a5d7 service nova] Lock "c7d08234-fcef-4cdb-939f-05ea9e8b08ea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.084871] env[62952]: DEBUG oslo_concurrency.lockutils [req-1f01050d-3316-4f41-95fe-92cb359b8988 req-a4c79c27-31d7-4de9-9db0-271be032a5d7 service nova] Lock "c7d08234-fcef-4cdb-939f-05ea9e8b08ea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.085046] env[62952]: DEBUG nova.compute.manager [req-1f01050d-3316-4f41-95fe-92cb359b8988 req-a4c79c27-31d7-4de9-9db0-271be032a5d7 service nova] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] No waiting events found dispatching network-vif-plugged-89ebaa6b-2763-41e0-a0d7-954adf9ac031 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 972.085369] env[62952]: WARNING nova.compute.manager [req-1f01050d-3316-4f41-95fe-92cb359b8988 req-a4c79c27-31d7-4de9-9db0-271be032a5d7 service nova] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Received unexpected event network-vif-plugged-89ebaa6b-2763-41e0-a0d7-954adf9ac031 for instance with vm_state building and task_state spawning. [ 972.187217] env[62952]: DEBUG nova.network.neutron [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Successfully updated port: 89ebaa6b-2763-41e0-a0d7-954adf9ac031 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 972.218792] env[62952]: DEBUG oslo_vmware.api [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263530, 'name': ReconfigVM_Task, 'duration_secs': 0.232382} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.218996] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Reconfigured VM instance instance-0000004c to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 972.223692] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc8fc51e-e9a2-4453-8480-d79a001aa903 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.242974] env[62952]: DEBUG oslo_vmware.api [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 972.242974] env[62952]: value = "task-1263532" [ 972.242974] env[62952]: _type = "Task" [ 972.242974] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.253769] env[62952]: DEBUG oslo_vmware.api [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263532, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.266547] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8420de1e-6de3-4cfd-bf1c-5293b4e52c9f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.275911] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a332f11a-2a62-4dac-8e73-9619b47bce39 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.312047] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.312274] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.313479] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e90cd6f-b116-4dc0-a0e5-d7788c345999 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.318663] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d3b3f8-63b9-44a0-bb14-72d319f39697 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.343767] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ef553f-3fc8-43d7-8f29-7871a187488c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.348877] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f10e74-68bf-4230-bdb5-46ea16ce384e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.383042] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263531, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.383602] env[62952]: DEBUG nova.compute.provider_tree [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.391028] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Reconfiguring VM to detach interface {{(pid=62952) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 972.391752] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ee57113-95a1-446f-b2b3-84b9d9ce21c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.417507] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263528, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.419103] env[62952]: DEBUG oslo_vmware.api [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 972.419103] env[62952]: value = "task-1263533" [ 972.419103] env[62952]: _type = "Task" [ 972.419103] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.428027] env[62952]: DEBUG oslo_vmware.api [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263533, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.477816] env[62952]: DEBUG nova.compute.manager [req-a7da0e43-b06d-4f39-8588-8a04129c68f2 req-86207778-c14c-4891-996d-246636ad5497 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Received event network-vif-plugged-6a5ee171-10c8-43aa-8b0b-12f2d7068e6e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 972.478087] env[62952]: DEBUG oslo_concurrency.lockutils [req-a7da0e43-b06d-4f39-8588-8a04129c68f2 req-86207778-c14c-4891-996d-246636ad5497 service nova] Acquiring lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.478417] env[62952]: DEBUG oslo_concurrency.lockutils [req-a7da0e43-b06d-4f39-8588-8a04129c68f2 req-86207778-c14c-4891-996d-246636ad5497 service nova] Lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.478539] env[62952]: DEBUG oslo_concurrency.lockutils [req-a7da0e43-b06d-4f39-8588-8a04129c68f2 req-86207778-c14c-4891-996d-246636ad5497 service nova] Lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.478717] env[62952]: DEBUG nova.compute.manager [req-a7da0e43-b06d-4f39-8588-8a04129c68f2 req-86207778-c14c-4891-996d-246636ad5497 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] No waiting events found dispatching network-vif-plugged-6a5ee171-10c8-43aa-8b0b-12f2d7068e6e {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 972.478888] env[62952]: WARNING nova.compute.manager [req-a7da0e43-b06d-4f39-8588-8a04129c68f2 req-86207778-c14c-4891-996d-246636ad5497 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Received unexpected event network-vif-plugged-6a5ee171-10c8-43aa-8b0b-12f2d7068e6e for instance with vm_state building and task_state spawning. [ 972.479076] env[62952]: DEBUG nova.compute.manager [req-a7da0e43-b06d-4f39-8588-8a04129c68f2 req-86207778-c14c-4891-996d-246636ad5497 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Received event network-changed-6a5ee171-10c8-43aa-8b0b-12f2d7068e6e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 972.479327] env[62952]: DEBUG nova.compute.manager [req-a7da0e43-b06d-4f39-8588-8a04129c68f2 req-86207778-c14c-4891-996d-246636ad5497 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Refreshing instance network info cache due to event network-changed-6a5ee171-10c8-43aa-8b0b-12f2d7068e6e. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 972.479504] env[62952]: DEBUG oslo_concurrency.lockutils [req-a7da0e43-b06d-4f39-8588-8a04129c68f2 req-86207778-c14c-4891-996d-246636ad5497 service nova] Acquiring lock "refresh_cache-fba2e955-8b5d-461c-997b-ef8f3ec0d906" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.479631] env[62952]: DEBUG oslo_concurrency.lockutils [req-a7da0e43-b06d-4f39-8588-8a04129c68f2 req-86207778-c14c-4891-996d-246636ad5497 service nova] Acquired lock "refresh_cache-fba2e955-8b5d-461c-997b-ef8f3ec0d906" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.479760] env[62952]: DEBUG nova.network.neutron [req-a7da0e43-b06d-4f39-8588-8a04129c68f2 req-86207778-c14c-4891-996d-246636ad5497 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Refreshing network info cache for port 6a5ee171-10c8-43aa-8b0b-12f2d7068e6e {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 972.689870] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "refresh_cache-c7d08234-fcef-4cdb-939f-05ea9e8b08ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.690271] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquired lock "refresh_cache-c7d08234-fcef-4cdb-939f-05ea9e8b08ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.690271] env[62952]: DEBUG nova.network.neutron [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 972.755333] env[62952]: DEBUG oslo_vmware.api [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263532, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.794217] env[62952]: DEBUG nova.network.neutron [None req-fcc259d7-b371-42aa-bfb0-c0446f0ea279 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance_info_cache with network_info: [{"id": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "address": "fa:16:3e:eb:c2:8d", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e2e9d1e-dc", "ovs_interfaceid": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.847754] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263531, 'name': PowerOffVM_Task, 'duration_secs': 1.017958} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.848084] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 972.848929] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365fa3cd-3dc9-4bc8-84ad-bf1d10fb827e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.868799] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ccf1f0-615d-44c7-91dc-5ed421cd66c8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.892770] env[62952]: DEBUG nova.scheduler.client.report [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 972.907556] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263528, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.910262] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 972.910262] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f0f3e363-d021-4fca-87bf-854013ce2a2a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.917451] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 972.917451] env[62952]: value = "task-1263534" [ 972.917451] env[62952]: _type = "Task" [ 972.917451] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.933432] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 972.933666] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 972.933921] env[62952]: DEBUG oslo_concurrency.lockutils [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.934088] env[62952]: DEBUG oslo_concurrency.lockutils [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.934278] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 972.934552] env[62952]: DEBUG oslo_vmware.api [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.935267] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-daff41e3-7380-424f-bf29-23eb40f6129e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.950452] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 972.950693] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 972.951726] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-542d38fe-6a29-42d7-b1ac-652e08649ffc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.958162] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 972.958162] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a41ccd-62c9-ee19-2765-c85b165178ad" [ 972.958162] env[62952]: _type = "Task" [ 972.958162] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.966340] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a41ccd-62c9-ee19-2765-c85b165178ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.022036] env[62952]: DEBUG nova.network.neutron [req-a7da0e43-b06d-4f39-8588-8a04129c68f2 req-86207778-c14c-4891-996d-246636ad5497 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 973.161895] env[62952]: DEBUG nova.network.neutron [req-a7da0e43-b06d-4f39-8588-8a04129c68f2 req-86207778-c14c-4891-996d-246636ad5497 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.236473] env[62952]: DEBUG nova.network.neutron [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 973.252409] env[62952]: DEBUG oslo_vmware.api [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263532, 'name': ReconfigVM_Task, 'duration_secs': 0.818576} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.252695] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271922', 'volume_id': 'af7338aa-c509-4319-a3a4-ce7e8aea5cfe', 'name': 'volume-af7338aa-c509-4319-a3a4-ce7e8aea5cfe', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e', 'attached_at': '', 'detached_at': '', 'volume_id': 'af7338aa-c509-4319-a3a4-ce7e8aea5cfe', 'serial': 'af7338aa-c509-4319-a3a4-ce7e8aea5cfe'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 973.252968] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 973.253706] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29550d73-019b-49e8-9b91-eeee70863314 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.260307] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 973.262260] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7bb3f7d6-ad1a-4363-af5c-9ca9a85002d4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.297623] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fcc259d7-b371-42aa-bfb0-c0446f0ea279 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.301680] env[62952]: DEBUG nova.compute.manager [None req-fcc259d7-b371-42aa-bfb0-c0446f0ea279 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.302764] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-955e321e-e07e-4a5a-ae4b-bbec9c79d0bc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.328406] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 973.328653] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 973.328841] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleting the datastore file [datastore2] b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 973.329124] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d6d9c2e6-5e18-4580-9ac8-eec6c02cca11 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.336285] env[62952]: DEBUG oslo_vmware.api [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 973.336285] env[62952]: value = "task-1263536" [ 973.336285] env[62952]: _type = "Task" [ 973.336285] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.344921] env[62952]: DEBUG oslo_vmware.api [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263536, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.405720] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.901s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.407666] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263528, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.847828} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.408111] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.650s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.409737] env[62952]: INFO nova.virt.vmwareapi.ds_util [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_afa09349-4793-4e43-8ef6-1ebac150fc2a/OSTACK_IMG_afa09349-4793-4e43-8ef6-1ebac150fc2a.vmdk to [datastore1] devstack-image-cache_base/315e9b97-1fa4-4f98-9bd7-cd708e6717cd/315e9b97-1fa4-4f98-9bd7-cd708e6717cd.vmdk. [ 973.409941] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Cleaning up location [datastore1] OSTACK_IMG_afa09349-4793-4e43-8ef6-1ebac150fc2a {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 973.410132] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_afa09349-4793-4e43-8ef6-1ebac150fc2a {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 973.410605] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1b1651fb-020c-441e-968b-8490b58c6b30 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.417142] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 973.417142] env[62952]: value = "task-1263537" [ 973.417142] env[62952]: _type = "Task" [ 973.417142] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.428179] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263537, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.433195] env[62952]: DEBUG oslo_vmware.api [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.434135] env[62952]: INFO nova.scheduler.client.report [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleted allocations for instance 20ea323b-8961-4981-8045-f9cb85815d99 [ 973.435736] env[62952]: DEBUG nova.network.neutron [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Updating instance_info_cache with network_info: [{"id": "89ebaa6b-2763-41e0-a0d7-954adf9ac031", "address": "fa:16:3e:a5:fa:e1", "network": {"id": "2c5146eb-dcff-470d-984f-b96a3c91bb3e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1950178234-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1487c06883e444a589d9b688155dc00e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89ebaa6b-27", "ovs_interfaceid": "89ebaa6b-2763-41e0-a0d7-954adf9ac031", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.470925] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a41ccd-62c9-ee19-2765-c85b165178ad, 'name': SearchDatastore_Task, 'duration_secs': 0.081348} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.471753] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed11b811-7f12-4f64-bb31-72624281b993 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.478036] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 973.478036] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5233a816-c009-14c1-0bbc-864510e6ee52" [ 973.478036] env[62952]: _type = "Task" [ 973.478036] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.486948] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5233a816-c009-14c1-0bbc-864510e6ee52, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.665311] env[62952]: DEBUG oslo_concurrency.lockutils [req-a7da0e43-b06d-4f39-8588-8a04129c68f2 req-86207778-c14c-4891-996d-246636ad5497 service nova] Releasing lock "refresh_cache-fba2e955-8b5d-461c-997b-ef8f3ec0d906" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.702613] env[62952]: DEBUG nova.network.neutron [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Successfully updated port: b0b3ad2c-5b56-4239-856d-663404db8324 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 973.847604] env[62952]: DEBUG oslo_vmware.api [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263536, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1448} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.848499] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.848779] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 973.848895] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 973.849060] env[62952]: INFO nova.compute.manager [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Took 2.76 seconds to destroy the instance on the hypervisor. [ 973.849318] env[62952]: DEBUG oslo.service.loopingcall [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.849513] env[62952]: DEBUG nova.compute.manager [-] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 973.849609] env[62952]: DEBUG nova.network.neutron [-] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 973.939835] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263537, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.034323} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.947029] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Releasing lock "refresh_cache-c7d08234-fcef-4cdb-939f-05ea9e8b08ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.947029] env[62952]: DEBUG nova.compute.manager [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Instance network_info: |[{"id": "89ebaa6b-2763-41e0-a0d7-954adf9ac031", "address": "fa:16:3e:a5:fa:e1", "network": {"id": "2c5146eb-dcff-470d-984f-b96a3c91bb3e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1950178234-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1487c06883e444a589d9b688155dc00e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89ebaa6b-27", "ovs_interfaceid": "89ebaa6b-2763-41e0-a0d7-954adf9ac031", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 973.947029] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.947029] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Releasing lock "[datastore1] devstack-image-cache_base/315e9b97-1fa4-4f98-9bd7-cd708e6717cd/315e9b97-1fa4-4f98-9bd7-cd708e6717cd.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.947029] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/315e9b97-1fa4-4f98-9bd7-cd708e6717cd/315e9b97-1fa4-4f98-9bd7-cd708e6717cd.vmdk to [datastore1] 9cc7bc2f-ff22-43cf-be79-ff8c6426115b/9cc7bc2f-ff22-43cf-be79-ff8c6426115b.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 973.950115] env[62952]: DEBUG oslo_vmware.api [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.950733] env[62952]: DEBUG oslo_concurrency.lockutils [None req-95d1ff39-6d27-409a-bf82-c7beaaaa5d6e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "20ea323b-8961-4981-8045-f9cb85815d99" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.049s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.952123] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:fa:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e445fb59-822c-4d7d-943b-c8e3bbaca62e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '89ebaa6b-2763-41e0-a0d7-954adf9ac031', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 973.965275] env[62952]: DEBUG oslo.service.loopingcall [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.965608] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c9577e9-a1bd-4b26-824a-7ea8bebeb561 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.971018] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 973.971340] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ecfeeeb7-b463-47ea-88f6-f0c570169d6d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.007962] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 974.007962] env[62952]: value = "task-1263538" [ 974.007962] env[62952]: _type = "Task" [ 974.007962] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.012641] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5233a816-c009-14c1-0bbc-864510e6ee52, 'name': SearchDatastore_Task, 'duration_secs': 0.013421} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.012837] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 974.012837] env[62952]: value = "task-1263539" [ 974.012837] env[62952]: _type = "Task" [ 974.012837] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.015833] env[62952]: DEBUG oslo_concurrency.lockutils [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.016130] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 77fe79de-8da8-486e-af63-91f9d8196ca9/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk. {{(pid=62952) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 974.016441] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-04a0a1dd-801e-4d56-b821-1569b4fd5d40 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.025929] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263538, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.029944] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 974.029944] env[62952]: value = "task-1263540" [ 974.029944] env[62952]: _type = "Task" [ 974.029944] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.030214] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263539, 'name': CreateVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.039352] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263540, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.128248] env[62952]: DEBUG nova.compute.manager [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Received event network-changed-89ebaa6b-2763-41e0-a0d7-954adf9ac031 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 974.128608] env[62952]: DEBUG nova.compute.manager [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Refreshing instance network info cache due to event network-changed-89ebaa6b-2763-41e0-a0d7-954adf9ac031. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 974.128886] env[62952]: DEBUG oslo_concurrency.lockutils [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] Acquiring lock "refresh_cache-c7d08234-fcef-4cdb-939f-05ea9e8b08ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.129045] env[62952]: DEBUG oslo_concurrency.lockutils [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] Acquired lock "refresh_cache-c7d08234-fcef-4cdb-939f-05ea9e8b08ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.129229] env[62952]: DEBUG nova.network.neutron [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Refreshing network info cache for port 89ebaa6b-2763-41e0-a0d7-954adf9ac031 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 974.208056] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquiring lock "refresh_cache-fba2e955-8b5d-461c-997b-ef8f3ec0d906" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.208056] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquired lock "refresh_cache-fba2e955-8b5d-461c-997b-ef8f3ec0d906" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.208056] env[62952]: DEBUG nova.network.neutron [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 974.318494] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5518300e-a8be-4d0e-9b5c-a4f304586f68 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.326432] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fcc259d7-b371-42aa-bfb0-c0446f0ea279 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Doing hard reboot of VM {{(pid=62952) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 974.326746] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-6d360119-aef6-41fd-b809-8b5e1f1c4f97 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.336087] env[62952]: DEBUG oslo_vmware.api [None req-fcc259d7-b371-42aa-bfb0-c0446f0ea279 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 974.336087] env[62952]: value = "task-1263541" [ 974.336087] env[62952]: _type = "Task" [ 974.336087] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.344221] env[62952]: DEBUG oslo_vmware.api [None req-fcc259d7-b371-42aa-bfb0-c0446f0ea279 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263541, 'name': ResetVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.443769] env[62952]: DEBUG oslo_vmware.api [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.457182] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance f4f8bee4-bf92-4321-afdc-16f7fb17dbae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 974.457348] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 0533ee81-b69d-4f4e-9832-78c5ab4512b1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 974.457476] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 6e289c20-8dd2-4680-b816-9bbe82fe7aec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 974.457595] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 14f8dbbe-2c4a-4948-bf15-106d5b1b0677 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 974.457736] env[62952]: WARNING nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 974.457852] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance dafa4926-3ab5-47f9-8ff2-4049e05c3113 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 974.457978] env[62952]: WARNING nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance e17b3d83-7cea-4a65-9457-7e57a6e7641a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 974.458109] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 9cc7bc2f-ff22-43cf-be79-ff8c6426115b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 974.458226] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 974.458341] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance dbca4369-1ed3-493a-9847-9fa1e4293475 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 974.458463] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 77fe79de-8da8-486e-af63-91f9d8196ca9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 974.458577] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance fba2e955-8b5d-461c-997b-ef8f3ec0d906 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 974.458691] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance c7d08234-fcef-4cdb-939f-05ea9e8b08ea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 974.522955] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263538, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.530397] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263539, 'name': CreateVM_Task, 'duration_secs': 0.467119} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.530681] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 974.534308] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.534491] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.534836] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 974.535160] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee31b100-9f87-4314-b5cc-527de97da39d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.541916] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 974.541916] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f08296-96ae-23a3-2a4a-1a73d85db6d9" [ 974.541916] env[62952]: _type = "Task" [ 974.541916] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.546707] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263540, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.556611] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f08296-96ae-23a3-2a4a-1a73d85db6d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.601197] env[62952]: DEBUG nova.network.neutron [-] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.650103] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "6f0b04f5-811f-4c53-808b-6d9d22100a86" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.650412] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "6f0b04f5-811f-4c53-808b-6d9d22100a86" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.773830] env[62952]: DEBUG nova.network.neutron [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 974.850371] env[62952]: DEBUG oslo_vmware.api [None req-fcc259d7-b371-42aa-bfb0-c0446f0ea279 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263541, 'name': ResetVM_Task, 'duration_secs': 0.111797} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.850965] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fcc259d7-b371-42aa-bfb0-c0446f0ea279 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Did hard reboot of VM {{(pid=62952) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 974.851100] env[62952]: DEBUG nova.compute.manager [None req-fcc259d7-b371-42aa-bfb0-c0446f0ea279 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 974.852017] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684d98a4-9bb0-4308-98e9-41c3c5028b2f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.893790] env[62952]: DEBUG nova.network.neutron [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Updated VIF entry in instance network info cache for port 89ebaa6b-2763-41e0-a0d7-954adf9ac031. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 974.894234] env[62952]: DEBUG nova.network.neutron [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Updating instance_info_cache with network_info: [{"id": "89ebaa6b-2763-41e0-a0d7-954adf9ac031", "address": "fa:16:3e:a5:fa:e1", "network": {"id": "2c5146eb-dcff-470d-984f-b96a3c91bb3e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1950178234-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1487c06883e444a589d9b688155dc00e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89ebaa6b-27", "ovs_interfaceid": "89ebaa6b-2763-41e0-a0d7-954adf9ac031", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.948800] env[62952]: DEBUG oslo_vmware.api [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.962121] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 975.025484] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263538, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.043799] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263540, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.059275] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f08296-96ae-23a3-2a4a-1a73d85db6d9, 'name': SearchDatastore_Task, 'duration_secs': 0.109425} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.059759] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.060097] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 975.060357] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.060503] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.060687] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 975.061070] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f2dde704-6320-425c-9e18-b1974f40ab0c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.084558] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 975.084872] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 975.089183] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98aabfde-b9a1-427c-8e7a-ac0f7812bba2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.097924] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 975.097924] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e75bf0-5c5f-55a7-b8d9-bccde09da728" [ 975.097924] env[62952]: _type = "Task" [ 975.097924] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.104756] env[62952]: INFO nova.compute.manager [-] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Took 1.25 seconds to deallocate network for instance. [ 975.111917] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e75bf0-5c5f-55a7-b8d9-bccde09da728, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.153203] env[62952]: DEBUG nova.compute.manager [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 975.369086] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fcc259d7-b371-42aa-bfb0-c0446f0ea279 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.950s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.396981] env[62952]: DEBUG oslo_concurrency.lockutils [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] Releasing lock "refresh_cache-c7d08234-fcef-4cdb-939f-05ea9e8b08ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.397168] env[62952]: DEBUG nova.compute.manager [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Received event network-vif-plugged-b0b3ad2c-5b56-4239-856d-663404db8324 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 975.397403] env[62952]: DEBUG oslo_concurrency.lockutils [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] Acquiring lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.397640] env[62952]: DEBUG oslo_concurrency.lockutils [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] Lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.398010] env[62952]: DEBUG oslo_concurrency.lockutils [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] Lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.398248] env[62952]: DEBUG nova.compute.manager [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] No waiting events found dispatching network-vif-plugged-b0b3ad2c-5b56-4239-856d-663404db8324 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 975.398448] env[62952]: WARNING nova.compute.manager [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Received unexpected event network-vif-plugged-b0b3ad2c-5b56-4239-856d-663404db8324 for instance with vm_state building and task_state spawning. [ 975.398625] env[62952]: DEBUG nova.compute.manager [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Received event network-changed-b0b3ad2c-5b56-4239-856d-663404db8324 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 975.398884] env[62952]: DEBUG nova.compute.manager [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Refreshing instance network info cache due to event network-changed-b0b3ad2c-5b56-4239-856d-663404db8324. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 975.399194] env[62952]: DEBUG oslo_concurrency.lockutils [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] Acquiring lock "refresh_cache-fba2e955-8b5d-461c-997b-ef8f3ec0d906" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.442953] env[62952]: DEBUG nova.network.neutron [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Updating instance_info_cache with network_info: [{"id": "34b7738d-65b4-4571-bef9-bf3c5ff297ef", "address": "fa:16:3e:08:24:10", "network": {"id": "4a13ff4b-822f-4bbf-bdaf-8ae41cc3ad82", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-762689622", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.144", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26f601d669ce42dca2f875ca3a68525c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34b7738d-65", "ovs_interfaceid": "34b7738d-65b4-4571-bef9-bf3c5ff297ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6a5ee171-10c8-43aa-8b0b-12f2d7068e6e", "address": "fa:16:3e:90:c0:2b", "network": {"id": "0b781d47-e39c-4553-b001-9ad42653a08e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-532489602", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.244", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "26f601d669ce42dca2f875ca3a68525c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "305ccd93-08cb-4658-845c-d9b64952daf7", "external-id": "nsx-vlan-transportzone-490", "segmentation_id": 490, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a5ee171-10", "ovs_interfaceid": "6a5ee171-10c8-43aa-8b0b-12f2d7068e6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b0b3ad2c-5b56-4239-856d-663404db8324", "address": "fa:16:3e:ff:19:6d", "network": {"id": "4a13ff4b-822f-4bbf-bdaf-8ae41cc3ad82", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-762689622", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26f601d669ce42dca2f875ca3a68525c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0b3ad2c-5b", "ovs_interfaceid": "b0b3ad2c-5b56-4239-856d-663404db8324", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.452899] env[62952]: DEBUG oslo_vmware.api [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.465385] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 7ba1587d-edaf-4dce-a224-6b2ac22cfecd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 975.525103] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263538, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.543760] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263540, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.212823} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.543760] env[62952]: INFO nova.virt.vmwareapi.ds_util [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 77fe79de-8da8-486e-af63-91f9d8196ca9/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk. [ 975.544581] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee221bc-d40e-402a-913b-6803bce2256f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.573434] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 77fe79de-8da8-486e-af63-91f9d8196ca9/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 975.573808] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71dd2373-2f0a-495e-a3ae-ddcc208e249a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.594887] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 975.594887] env[62952]: value = "task-1263542" [ 975.594887] env[62952]: _type = "Task" [ 975.594887] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.608644] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e75bf0-5c5f-55a7-b8d9-bccde09da728, 'name': SearchDatastore_Task, 'duration_secs': 0.089238} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.612858] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263542, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.613117] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8bee985-ba91-4509-b963-f3a0c4bc081c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.620867] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 975.620867] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5243bf9e-e6ad-d80b-98f2-df74a0a630eb" [ 975.620867] env[62952]: _type = "Task" [ 975.620867] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.632405] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5243bf9e-e6ad-d80b-98f2-df74a0a630eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.667307] env[62952]: INFO nova.compute.manager [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Took 0.56 seconds to detach 1 volumes for instance. [ 975.681062] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.946029] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Releasing lock "refresh_cache-fba2e955-8b5d-461c-997b-ef8f3ec0d906" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.946836] env[62952]: DEBUG nova.compute.manager [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Instance network_info: |[{"id": "34b7738d-65b4-4571-bef9-bf3c5ff297ef", "address": "fa:16:3e:08:24:10", "network": {"id": "4a13ff4b-822f-4bbf-bdaf-8ae41cc3ad82", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-762689622", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.144", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26f601d669ce42dca2f875ca3a68525c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34b7738d-65", "ovs_interfaceid": "34b7738d-65b4-4571-bef9-bf3c5ff297ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6a5ee171-10c8-43aa-8b0b-12f2d7068e6e", "address": "fa:16:3e:90:c0:2b", "network": {"id": "0b781d47-e39c-4553-b001-9ad42653a08e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-532489602", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.244", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "26f601d669ce42dca2f875ca3a68525c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "305ccd93-08cb-4658-845c-d9b64952daf7", "external-id": "nsx-vlan-transportzone-490", "segmentation_id": 490, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a5ee171-10", "ovs_interfaceid": "6a5ee171-10c8-43aa-8b0b-12f2d7068e6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b0b3ad2c-5b56-4239-856d-663404db8324", "address": "fa:16:3e:ff:19:6d", "network": {"id": "4a13ff4b-822f-4bbf-bdaf-8ae41cc3ad82", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-762689622", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26f601d669ce42dca2f875ca3a68525c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0b3ad2c-5b", "ovs_interfaceid": "b0b3ad2c-5b56-4239-856d-663404db8324", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 975.950393] env[62952]: DEBUG oslo_concurrency.lockutils [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] Acquired lock "refresh_cache-fba2e955-8b5d-461c-997b-ef8f3ec0d906" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.950664] env[62952]: DEBUG nova.network.neutron [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Refreshing network info cache for port b0b3ad2c-5b56-4239-856d-663404db8324 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 975.952166] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:24:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '975b168a-03e5-449d-95ac-4d51ba027242', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '34b7738d-65b4-4571-bef9-bf3c5ff297ef', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:c0:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '305ccd93-08cb-4658-845c-d9b64952daf7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6a5ee171-10c8-43aa-8b0b-12f2d7068e6e', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:ff:19:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '975b168a-03e5-449d-95ac-4d51ba027242', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b0b3ad2c-5b56-4239-856d-663404db8324', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 975.963211] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Creating folder: Project (26f601d669ce42dca2f875ca3a68525c). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 975.963556] env[62952]: DEBUG oslo_vmware.api [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.966881] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8d0b9a8a-4851-47c1-868c-96bcccfe0a28 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.969744] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 6f0b04f5-811f-4c53-808b-6d9d22100a86 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 975.969990] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 975.970157] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2624MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 975.987450] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Created folder: Project (26f601d669ce42dca2f875ca3a68525c) in parent group-v271811. [ 975.987684] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Creating folder: Instances. Parent ref: group-v271926. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 975.987965] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cd5b2f4d-b9c9-426c-930f-f27d9f97ec82 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.007527] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Created folder: Instances in parent group-v271926. [ 976.007830] env[62952]: DEBUG oslo.service.loopingcall [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.008065] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 976.008338] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-edf3495e-0d67-4823-b2f1-110a1af3c243 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.043122] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263538, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.044696] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 976.044696] env[62952]: value = "task-1263545" [ 976.044696] env[62952]: _type = "Task" [ 976.044696] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.065866] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263545, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.111227] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263542, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.133832] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5243bf9e-e6ad-d80b-98f2-df74a0a630eb, 'name': SearchDatastore_Task, 'duration_secs': 0.090215} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.133832] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.134055] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] c7d08234-fcef-4cdb-939f-05ea9e8b08ea/c7d08234-fcef-4cdb-939f-05ea9e8b08ea.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 976.134379] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5a3f178d-ff60-45ff-8d2d-8c052101f5a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.145744] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 976.145744] env[62952]: value = "task-1263546" [ 976.145744] env[62952]: _type = "Task" [ 976.145744] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.159390] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263546, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.173869] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.239745] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac4c44cb-9044-4ab7-b703-6c670e5f84d0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.252698] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84441d50-af3e-4baa-b062-74b406f27be7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.285980] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c6423c8-e134-4400-bf21-33e94c354c42 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.297158] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d5f80f0-0c94-4bad-9da7-0757770e76dd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.314873] env[62952]: DEBUG nova.compute.provider_tree [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.342463] env[62952]: DEBUG nova.compute.manager [req-43951750-843f-4507-8a1e-000f4bcc471b req-84b95166-9436-46d4-83d0-70905abaeabc service nova] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Received event network-vif-deleted-b708a98e-0fd2-4730-8307-8731609c4aa3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 976.452219] env[62952]: DEBUG oslo_vmware.api [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.544639] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263538, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.557410] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263545, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.567244] env[62952]: DEBUG oslo_concurrency.lockutils [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.567609] env[62952]: DEBUG oslo_concurrency.lockutils [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.567850] env[62952]: DEBUG oslo_concurrency.lockutils [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.568136] env[62952]: DEBUG oslo_concurrency.lockutils [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.568346] env[62952]: DEBUG oslo_concurrency.lockutils [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.570963] env[62952]: INFO nova.compute.manager [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Terminating instance [ 976.572676] env[62952]: DEBUG nova.compute.manager [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 976.572886] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 976.573788] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c03b112-0661-4109-8c09-23c222b5de69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.587182] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 976.587182] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9bfe4ac-e1c9-493b-80ca-b63ff5e57714 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.596564] env[62952]: DEBUG oslo_vmware.api [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 976.596564] env[62952]: value = "task-1263547" [ 976.596564] env[62952]: _type = "Task" [ 976.596564] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.609803] env[62952]: DEBUG oslo_vmware.api [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263547, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.613391] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263542, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.628971] env[62952]: DEBUG nova.network.neutron [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Updated VIF entry in instance network info cache for port b0b3ad2c-5b56-4239-856d-663404db8324. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 976.629507] env[62952]: DEBUG nova.network.neutron [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Updating instance_info_cache with network_info: [{"id": "34b7738d-65b4-4571-bef9-bf3c5ff297ef", "address": "fa:16:3e:08:24:10", "network": {"id": "4a13ff4b-822f-4bbf-bdaf-8ae41cc3ad82", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-762689622", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.144", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26f601d669ce42dca2f875ca3a68525c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34b7738d-65", "ovs_interfaceid": "34b7738d-65b4-4571-bef9-bf3c5ff297ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6a5ee171-10c8-43aa-8b0b-12f2d7068e6e", "address": "fa:16:3e:90:c0:2b", "network": {"id": "0b781d47-e39c-4553-b001-9ad42653a08e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-532489602", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.244", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "26f601d669ce42dca2f875ca3a68525c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "305ccd93-08cb-4658-845c-d9b64952daf7", "external-id": "nsx-vlan-transportzone-490", "segmentation_id": 490, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a5ee171-10", "ovs_interfaceid": "6a5ee171-10c8-43aa-8b0b-12f2d7068e6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b0b3ad2c-5b56-4239-856d-663404db8324", "address": "fa:16:3e:ff:19:6d", "network": {"id": "4a13ff4b-822f-4bbf-bdaf-8ae41cc3ad82", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-762689622", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26f601d669ce42dca2f875ca3a68525c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0b3ad2c-5b", "ovs_interfaceid": "b0b3ad2c-5b56-4239-856d-663404db8324", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.658959] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263546, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.819305] env[62952]: DEBUG nova.scheduler.client.report [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.946761] env[62952]: DEBUG oslo_vmware.api [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.044167] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263538, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.922583} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.044492] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/315e9b97-1fa4-4f98-9bd7-cd708e6717cd/315e9b97-1fa4-4f98-9bd7-cd708e6717cd.vmdk to [datastore1] 9cc7bc2f-ff22-43cf-be79-ff8c6426115b/9cc7bc2f-ff22-43cf-be79-ff8c6426115b.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 977.045394] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7492a53-8302-40b4-981a-c6f673efdeb0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.075337] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] 9cc7bc2f-ff22-43cf-be79-ff8c6426115b/9cc7bc2f-ff22-43cf-be79-ff8c6426115b.vmdk or device None with type streamOptimized {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 977.075816] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263545, 'name': CreateVM_Task, 'duration_secs': 0.536674} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.075926] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f12e2f3-1479-41cc-a7a8-05c54c690799 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.090482] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 977.091395] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.091567] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.091900] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 977.092573] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5eb93f1-af2a-4290-a9bb-88177321d4e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.098147] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 977.098147] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c80929-90c3-2c8b-f54a-0c5d440034b9" [ 977.098147] env[62952]: _type = "Task" [ 977.098147] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.102217] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 977.102217] env[62952]: value = "task-1263548" [ 977.102217] env[62952]: _type = "Task" [ 977.102217] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.124664] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263542, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.131505] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c80929-90c3-2c8b-f54a-0c5d440034b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.137333] env[62952]: DEBUG oslo_concurrency.lockutils [req-ad754534-99ff-4d21-a34c-d0e9b0bb60d6 req-1a960549-7880-4848-935d-33f8e36a2dc2 service nova] Releasing lock "refresh_cache-fba2e955-8b5d-461c-997b-ef8f3ec0d906" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.137738] env[62952]: DEBUG oslo_vmware.api [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263547, 'name': PowerOffVM_Task, 'duration_secs': 0.268585} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.137996] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 977.138204] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 977.138837] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f85e3047-52dc-4489-8bfa-1e9e5f82105c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.143160] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.154228] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263546, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.217474] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 977.217754] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 977.217949] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Deleting the datastore file [datastore2] f4f8bee4-bf92-4321-afdc-16f7fb17dbae {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 977.218261] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0bf4b0f2-0b7d-4c85-9e0c-fd5d7b3c1f0c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.226386] env[62952]: DEBUG oslo_vmware.api [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 977.226386] env[62952]: value = "task-1263550" [ 977.226386] env[62952]: _type = "Task" [ 977.226386] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.236272] env[62952]: DEBUG oslo_vmware.api [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263550, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.324515] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 977.324695] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.917s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.325124] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.086s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.325320] env[62952]: DEBUG nova.objects.instance [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62952) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 977.447892] env[62952]: DEBUG oslo_vmware.api [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.616079] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263542, 'name': ReconfigVM_Task, 'duration_secs': 1.961702} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.616369] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c80929-90c3-2c8b-f54a-0c5d440034b9, 'name': SearchDatastore_Task, 'duration_secs': 0.032423} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.617110] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 77fe79de-8da8-486e-af63-91f9d8196ca9/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 977.617610] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.617891] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 977.618154] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.618359] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.618594] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 977.619489] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c824a841-06c0-4e9d-a914-72fa616c40f8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.623020] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-911b04b2-3e61-4c71-b4fd-4c76f288c1f7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.627858] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.658355] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6fdbd7ce-67ad-4e73-98d2-2f464bda1f5b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.668474] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 977.668694] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 977.669472] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78bbf89e-b045-458e-a6ea-8ff7932daf70 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.679490] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263546, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.373694} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.681360] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] c7d08234-fcef-4cdb-939f-05ea9e8b08ea/c7d08234-fcef-4cdb-939f-05ea9e8b08ea.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 977.681591] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 977.681921] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 977.681921] env[62952]: value = "task-1263551" [ 977.681921] env[62952]: _type = "Task" [ 977.681921] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.682174] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 977.682174] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52045f82-379d-0d6b-e293-9a9fcc391ac8" [ 977.682174] env[62952]: _type = "Task" [ 977.682174] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.682360] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fae634cb-5dba-4c3c-b674-c6bea2446d14 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.700441] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52045f82-379d-0d6b-e293-9a9fcc391ac8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.700799] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263551, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.701815] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 977.701815] env[62952]: value = "task-1263552" [ 977.701815] env[62952]: _type = "Task" [ 977.701815] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.711554] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263552, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.737098] env[62952]: DEBUG oslo_vmware.api [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263550, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.948204] env[62952]: DEBUG oslo_vmware.api [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263533, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.117907] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.198169] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52045f82-379d-0d6b-e293-9a9fcc391ac8, 'name': SearchDatastore_Task, 'duration_secs': 0.034716} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.202113] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263551, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.202369] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e44fecde-c307-4762-9822-00faf45c1b07 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.212715] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263552, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078454} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.214007] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 978.214372] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 978.214372] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b4cdb4-0304-45ad-e3c9-9c6cf95a2de4" [ 978.214372] env[62952]: _type = "Task" [ 978.214372] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.215157] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557a0558-8306-4858-bebd-40214a759b66 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.225460] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b4cdb4-0304-45ad-e3c9-9c6cf95a2de4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.242941] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] c7d08234-fcef-4cdb-939f-05ea9e8b08ea/c7d08234-fcef-4cdb-939f-05ea9e8b08ea.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.245979] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f50c0207-8aac-44b5-95d1-abc318a918a0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.265613] env[62952]: DEBUG oslo_vmware.api [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263550, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.515659} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.266958] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 978.267163] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 978.267380] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 978.267576] env[62952]: INFO nova.compute.manager [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Took 1.69 seconds to destroy the instance on the hypervisor. [ 978.267852] env[62952]: DEBUG oslo.service.loopingcall [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.268129] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 978.268129] env[62952]: value = "task-1263553" [ 978.268129] env[62952]: _type = "Task" [ 978.268129] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.268318] env[62952]: DEBUG nova.compute.manager [-] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 978.268466] env[62952]: DEBUG nova.network.neutron [-] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 978.278576] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263553, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.337169] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7dfacd03-bbc3-40be-abf1-4814b488543f tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.338562] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.537s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.340168] env[62952]: INFO nova.compute.claims [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 978.449220] env[62952]: DEBUG oslo_vmware.api [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263533, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.618778] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263548, 'name': ReconfigVM_Task, 'duration_secs': 1.146281} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.619155] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Reconfigured VM instance instance-00000033 to attach disk [datastore1] 9cc7bc2f-ff22-43cf-be79-ff8c6426115b/9cc7bc2f-ff22-43cf-be79-ff8c6426115b.vmdk or device None with type streamOptimized {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.620606] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encryption_format': None, 'boot_index': 0, 'device_type': 'disk', 'device_name': '/dev/sda', 'size': 0, 'disk_bus': None, 'guest_format': None, 'encryption_options': None, 'encrypted': False, 'encryption_secret_uuid': None, 'image_id': '119dd2b4-b6de-42d5-adba-db818e4c2cd7'}], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271910', 'volume_id': '11fed6c5-a8bb-4173-ade8-22ca4aa0b85c', 'name': 'volume-11fed6c5-a8bb-4173-ade8-22ca4aa0b85c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '9cc7bc2f-ff22-43cf-be79-ff8c6426115b', 'attached_at': '', 'detached_at': '', 'volume_id': '11fed6c5-a8bb-4173-ade8-22ca4aa0b85c', 'serial': '11fed6c5-a8bb-4173-ade8-22ca4aa0b85c'}, 'boot_index': None, 'mount_device': '/dev/sdb', 'device_type': None, 'delete_on_termination': False, 'attachment_id': '302ecbb0-c1fd-465a-be1b-407a8b8594bb', 'disk_bus': None, 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=62952) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 978.620821] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 978.621033] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271910', 'volume_id': '11fed6c5-a8bb-4173-ade8-22ca4aa0b85c', 'name': 'volume-11fed6c5-a8bb-4173-ade8-22ca4aa0b85c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '9cc7bc2f-ff22-43cf-be79-ff8c6426115b', 'attached_at': '', 'detached_at': '', 'volume_id': '11fed6c5-a8bb-4173-ade8-22ca4aa0b85c', 'serial': '11fed6c5-a8bb-4173-ade8-22ca4aa0b85c'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 978.621924] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab83a331-b259-499c-b253-a92142b8435d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.639403] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e949ad4-f105-4bee-a7f9-cc40410b89e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.668271] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] volume-11fed6c5-a8bb-4173-ade8-22ca4aa0b85c/volume-11fed6c5-a8bb-4173-ade8-22ca4aa0b85c.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.668506] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45a50470-b6dd-48c3-a371-5a9bb005d266 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.690722] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 978.690722] env[62952]: value = "task-1263554" [ 978.690722] env[62952]: _type = "Task" [ 978.690722] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.697594] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263551, 'name': ReconfigVM_Task, 'duration_secs': 0.555528} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.698355] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 978.698725] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf0fd24c-abd8-4718-9e86-4f9875de5087 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.704057] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263554, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.708809] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 978.708809] env[62952]: value = "task-1263555" [ 978.708809] env[62952]: _type = "Task" [ 978.708809] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.717764] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263555, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.725959] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b4cdb4-0304-45ad-e3c9-9c6cf95a2de4, 'name': SearchDatastore_Task, 'duration_secs': 0.047722} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.726237] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.726551] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] fba2e955-8b5d-461c-997b-ef8f3ec0d906/fba2e955-8b5d-461c-997b-ef8f3ec0d906.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 978.726827] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-23e8f87e-563c-40f8-9b9a-dc4c8d2fd438 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.734756] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 978.734756] env[62952]: value = "task-1263556" [ 978.734756] env[62952]: _type = "Task" [ 978.734756] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.744724] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263556, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.782575] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263553, 'name': ReconfigVM_Task, 'duration_secs': 0.311875} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.782982] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Reconfigured VM instance instance-00000050 to attach disk [datastore1] c7d08234-fcef-4cdb-939f-05ea9e8b08ea/c7d08234-fcef-4cdb-939f-05ea9e8b08ea.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.783805] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5470b3b4-a538-45b3-8acc-ef11663a28d6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.793865] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 978.793865] env[62952]: value = "task-1263557" [ 978.793865] env[62952]: _type = "Task" [ 978.793865] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.804918] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263557, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.887768] env[62952]: DEBUG nova.compute.manager [req-823a3368-80c5-4ec8-a5e9-29db6d762181 req-f0476868-788f-46fc-938e-06dc13732085 service nova] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Received event network-vif-deleted-363ca83f-c6ac-4446-996c-edc459eb29bf {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 978.887906] env[62952]: INFO nova.compute.manager [req-823a3368-80c5-4ec8-a5e9-29db6d762181 req-f0476868-788f-46fc-938e-06dc13732085 service nova] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Neutron deleted interface 363ca83f-c6ac-4446-996c-edc459eb29bf; detaching it from the instance and deleting it from the info cache [ 978.888106] env[62952]: DEBUG nova.network.neutron [req-823a3368-80c5-4ec8-a5e9-29db6d762181 req-f0476868-788f-46fc-938e-06dc13732085 service nova] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.953788] env[62952]: DEBUG oslo_vmware.api [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263533, 'name': ReconfigVM_Task, 'duration_secs': 6.266675} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.953788] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.953788] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Reconfigured VM to detach interface {{(pid=62952) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 979.206495] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263554, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.221609] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263555, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.246544] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263556, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.306471] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263557, 'name': Rename_Task, 'duration_secs': 0.193977} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.307566] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 979.307566] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c3f2b6f-d4b4-4a8a-a5f4-aea3f580b159 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.316074] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 979.316074] env[62952]: value = "task-1263558" [ 979.316074] env[62952]: _type = "Task" [ 979.316074] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.331612] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263558, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.363376] env[62952]: DEBUG nova.network.neutron [-] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.393833] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e9cabf9b-d605-4a4d-a6fa-6fe401157f0a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.404349] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf494ccd-f9e8-472c-a2ec-242e6b4ca6b5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.442332] env[62952]: DEBUG nova.compute.manager [req-823a3368-80c5-4ec8-a5e9-29db6d762181 req-f0476868-788f-46fc-938e-06dc13732085 service nova] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Detach interface failed, port_id=363ca83f-c6ac-4446-996c-edc459eb29bf, reason: Instance f4f8bee4-bf92-4321-afdc-16f7fb17dbae could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 979.613226] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53994d8a-3ae5-4c54-8dcb-d0eacd4a07a7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.622861] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3016a756-2de5-4b28-8ac2-aa5475886148 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.654409] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f76eee6a-973b-46d3-a496-1cb8a3d92e8a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.663958] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5e0349a-d8fc-4ea7-826e-4868d06b50e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.679220] env[62952]: DEBUG nova.compute.provider_tree [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 979.703034] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263554, 'name': ReconfigVM_Task, 'duration_secs': 0.669571} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.703034] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Reconfigured VM instance instance-00000033 to attach disk [datastore1] volume-11fed6c5-a8bb-4173-ade8-22ca4aa0b85c/volume-11fed6c5-a8bb-4173-ade8-22ca4aa0b85c.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 979.709234] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ac16659-a058-4cba-a1c6-a4f62c03d49a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.728228] env[62952]: DEBUG oslo_vmware.api [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263555, 'name': PowerOnVM_Task, 'duration_secs': 0.703389} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.733019] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 979.733019] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 979.733019] env[62952]: value = "task-1263559" [ 979.733019] env[62952]: _type = "Task" [ 979.733019] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.733019] env[62952]: DEBUG nova.compute.manager [None req-663daaee-344b-4a39-9417-299b4cc267dc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 979.733546] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb3fae1-6e36-4800-abf5-d9e79120f185 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.749611] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263559, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.754621] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263556, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51533} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.754936] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] fba2e955-8b5d-461c-997b-ef8f3ec0d906/fba2e955-8b5d-461c-997b-ef8f3ec0d906.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 979.755219] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 979.755476] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-64322be1-8078-493a-8b98-cfca03ec6676 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.764341] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 979.764341] env[62952]: value = "task-1263560" [ 979.764341] env[62952]: _type = "Task" [ 979.764341] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.776931] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263560, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.828048] env[62952]: DEBUG oslo_vmware.api [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263558, 'name': PowerOnVM_Task, 'duration_secs': 0.496266} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.828698] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 979.828841] env[62952]: INFO nova.compute.manager [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Took 8.13 seconds to spawn the instance on the hypervisor. [ 979.829266] env[62952]: DEBUG nova.compute.manager [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 979.831581] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d90b84-291e-40d8-8ec1-dabb040e557a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.867050] env[62952]: INFO nova.compute.manager [-] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Took 1.60 seconds to deallocate network for instance. [ 980.185271] env[62952]: DEBUG nova.scheduler.client.report [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.243578] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263559, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.274495] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263560, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081833} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.274764] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 980.275538] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd5427de-a7d7-483f-a21b-f418fbdf9c11 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.303386] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] fba2e955-8b5d-461c-997b-ef8f3ec0d906/fba2e955-8b5d-461c-997b-ef8f3ec0d906.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 980.303612] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee98eae6-a710-461d-a085-a161616d2ea4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.317789] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.318056] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.318286] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.318512] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.318739] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.320952] env[62952]: INFO nova.compute.manager [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Terminating instance [ 980.323064] env[62952]: DEBUG nova.compute.manager [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 980.323146] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 980.323909] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a0bee7-4a54-44f6-a906-a7022097c796 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.329102] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 980.329102] env[62952]: value = "task-1263561" [ 980.329102] env[62952]: _type = "Task" [ 980.329102] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.331263] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 980.334363] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-29dc8294-4d26-44e6-967f-7143039e219c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.341799] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263561, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.343165] env[62952]: DEBUG oslo_vmware.api [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 980.343165] env[62952]: value = "task-1263562" [ 980.343165] env[62952]: _type = "Task" [ 980.343165] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.357321] env[62952]: DEBUG oslo_vmware.api [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263562, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.360210] env[62952]: INFO nova.compute.manager [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Took 26.43 seconds to build instance. [ 980.373489] env[62952]: DEBUG oslo_concurrency.lockutils [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.504882] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.505098] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquired lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.505286] env[62952]: DEBUG nova.network.neutron [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 980.690678] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.352s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.691201] env[62952]: DEBUG nova.compute.manager [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 980.693731] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.959s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.695121] env[62952]: INFO nova.compute.claims [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 980.744526] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263559, 'name': ReconfigVM_Task, 'duration_secs': 0.525119} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.744842] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271910', 'volume_id': '11fed6c5-a8bb-4173-ade8-22ca4aa0b85c', 'name': 'volume-11fed6c5-a8bb-4173-ade8-22ca4aa0b85c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '9cc7bc2f-ff22-43cf-be79-ff8c6426115b', 'attached_at': '', 'detached_at': '', 'volume_id': '11fed6c5-a8bb-4173-ade8-22ca4aa0b85c', 'serial': '11fed6c5-a8bb-4173-ade8-22ca4aa0b85c'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 980.745441] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b3b9e1e5-922a-4f73-8dac-c29d94e333c7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.753230] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 980.753230] env[62952]: value = "task-1263563" [ 980.753230] env[62952]: _type = "Task" [ 980.753230] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.762620] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263563, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.844308] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263561, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.853794] env[62952]: DEBUG oslo_vmware.api [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263562, 'name': PowerOffVM_Task, 'duration_secs': 0.218718} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.854105] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 980.854282] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 980.854564] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6b562aa9-da96-47d5-8a1d-d9be3f3774dc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.862095] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3976be36-a47d-4ae2-9e9f-71b10223cc78 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "c7d08234-fcef-4cdb-939f-05ea9e8b08ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.943s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.926612] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 980.927113] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 980.927339] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Deleting the datastore file [datastore1] dafa4926-3ab5-47f9-8ff2-4049e05c3113 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 980.927767] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-edf5c2f9-d9d2-48fe-90a8-45d78f702fc2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.940123] env[62952]: DEBUG oslo_vmware.api [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 980.940123] env[62952]: value = "task-1263565" [ 980.940123] env[62952]: _type = "Task" [ 980.940123] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.950479] env[62952]: DEBUG oslo_vmware.api [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263565, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.199353] env[62952]: DEBUG nova.compute.utils [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 981.203830] env[62952]: DEBUG nova.compute.manager [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 981.204010] env[62952]: DEBUG nova.network.neutron [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 981.255923] env[62952]: DEBUG nova.policy [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3ec1aecc0cf64aaea875a4280b114874', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fbb5ccb4d615429facd6932fb55b04e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 981.269485] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263563, 'name': Rename_Task, 'duration_secs': 0.165101} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.269754] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 981.269996] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5dce7e8d-4afb-4740-91de-b6d712ca0191 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.277723] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 981.277723] env[62952]: value = "task-1263566" [ 981.277723] env[62952]: _type = "Task" [ 981.277723] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.286291] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263566, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.314038] env[62952]: INFO nova.network.neutron [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Port bdfdd199-b7c7-47fd-bcd7-15760750128b from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 981.314442] env[62952]: DEBUG nova.network.neutron [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Updating instance_info_cache with network_info: [{"id": "6bab534b-286c-4391-aeeb-fbb57655925a", "address": "fa:16:3e:e2:f4:40", "network": {"id": "52af4564-da88-4cc8-b51b-07855da20a7c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-748618263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91fde3ceb94649f698459dfc174bf835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "838c9497-35dd-415e-96c7-8dc21b0cd4b3", "external-id": "nsx-vlan-transportzone-530", "segmentation_id": 530, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bab534b-28", "ovs_interfaceid": "6bab534b-286c-4391-aeeb-fbb57655925a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.342967] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263561, 'name': ReconfigVM_Task, 'duration_secs': 0.639991} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.343347] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Reconfigured VM instance instance-0000004f to attach disk [datastore1] fba2e955-8b5d-461c-997b-ef8f3ec0d906/fba2e955-8b5d-461c-997b-ef8f3ec0d906.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 981.344149] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-88b872b7-d802-46e0-91c2-98a8fb229ecf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.352838] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 981.352838] env[62952]: value = "task-1263567" [ 981.352838] env[62952]: _type = "Task" [ 981.352838] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.366617] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263567, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.451684] env[62952]: DEBUG oslo_vmware.api [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263565, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.413396} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.451986] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 981.452368] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 981.452423] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 981.452710] env[62952]: INFO nova.compute.manager [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Took 1.13 seconds to destroy the instance on the hypervisor. [ 981.452979] env[62952]: DEBUG oslo.service.loopingcall [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.453241] env[62952]: DEBUG nova.compute.manager [-] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 981.453363] env[62952]: DEBUG nova.network.neutron [-] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 981.544831] env[62952]: DEBUG nova.network.neutron [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Successfully created port: ea5399c4-f18f-4c0b-82b1-062064db974b {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 981.704587] env[62952]: DEBUG nova.compute.manager [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 981.795836] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263566, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.817642] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Releasing lock "refresh_cache-dafa4926-3ab5-47f9-8ff2-4049e05c3113" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.866511] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263567, 'name': Rename_Task, 'duration_secs': 0.173095} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.866802] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 981.867185] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2075be0-e284-45c1-86e8-7d870082ed30 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.876241] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 981.876241] env[62952]: value = "task-1263568" [ 981.876241] env[62952]: _type = "Task" [ 981.876241] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.886045] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263568, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.982765] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f75680-3d09-42b5-9c80-d23dbb382868 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.993480] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16d861b5-6bd6-450c-805f-1fb3c6178bdf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.027988] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a10d83f-ec5b-4758-a7e5-a8f69754442b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.036480] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaacbf6e-0944-4d62-9e3f-a41e6bcd5a6e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.053085] env[62952]: DEBUG nova.compute.provider_tree [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.251971] env[62952]: DEBUG nova.compute.manager [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 982.253080] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92baef42-099b-4554-b292-70e85bf239ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.272711] env[62952]: INFO nova.compute.manager [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Rescuing [ 982.273092] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "refresh_cache-dbca4369-1ed3-493a-9847-9fa1e4293475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.273324] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired lock "refresh_cache-dbca4369-1ed3-493a-9847-9fa1e4293475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.273548] env[62952]: DEBUG nova.network.neutron [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 982.291559] env[62952]: DEBUG oslo_vmware.api [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263566, 'name': PowerOnVM_Task, 'duration_secs': 0.806865} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.291873] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 982.325171] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16117fa0-2ed7-4a9a-b4a1-6243c047a4c6 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "interface-dafa4926-3ab5-47f9-8ff2-4049e05c3113-bdfdd199-b7c7-47fd-bcd7-15760750128b" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.530s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.387844] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263568, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.541292] env[62952]: DEBUG nova.compute.manager [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 982.541292] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a1508dc-7079-4ca5-bf73-bf360c2a73c8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.556884] env[62952]: DEBUG nova.scheduler.client.report [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.586019] env[62952]: DEBUG nova.compute.manager [req-5552ea1d-760c-4416-98f5-9b40c916139c req-9d47f1b6-35f1-4056-9055-8c05a2328c87 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Received event network-vif-deleted-6bab534b-286c-4391-aeeb-fbb57655925a {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 982.586323] env[62952]: INFO nova.compute.manager [req-5552ea1d-760c-4416-98f5-9b40c916139c req-9d47f1b6-35f1-4056-9055-8c05a2328c87 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Neutron deleted interface 6bab534b-286c-4391-aeeb-fbb57655925a; detaching it from the instance and deleting it from the info cache [ 982.586525] env[62952]: DEBUG nova.network.neutron [req-5552ea1d-760c-4416-98f5-9b40c916139c req-9d47f1b6-35f1-4056-9055-8c05a2328c87 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.718733] env[62952]: DEBUG nova.compute.manager [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 982.744435] env[62952]: DEBUG nova.virt.hardware [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 982.745577] env[62952]: DEBUG nova.virt.hardware [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 982.745991] env[62952]: DEBUG nova.virt.hardware [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 982.746231] env[62952]: DEBUG nova.virt.hardware [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 982.746387] env[62952]: DEBUG nova.virt.hardware [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 982.746541] env[62952]: DEBUG nova.virt.hardware [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 982.746952] env[62952]: DEBUG nova.virt.hardware [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 982.747154] env[62952]: DEBUG nova.virt.hardware [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 982.747368] env[62952]: DEBUG nova.virt.hardware [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 982.747792] env[62952]: DEBUG nova.virt.hardware [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 982.747984] env[62952]: DEBUG nova.virt.hardware [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 982.749335] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00825ae1-4b29-476d-8734-c239ea98b179 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.760338] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edae0caa-ad67-4673-ab81-f74d4f51d8dc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.779189] env[62952]: INFO nova.compute.manager [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] instance snapshotting [ 982.784674] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f0d7cc5-a994-4218-bf94-687eeb4442a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.809081] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b654fd4-5379-4177-8971-1f3a68507942 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.891028] env[62952]: DEBUG oslo_vmware.api [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263568, 'name': PowerOnVM_Task, 'duration_secs': 0.751864} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.891028] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 982.891028] env[62952]: INFO nova.compute.manager [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Took 15.51 seconds to spawn the instance on the hypervisor. [ 982.891028] env[62952]: DEBUG nova.compute.manager [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 982.891028] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ba3d00d-f111-4d5b-97ba-c573be1ce51a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.956825] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46f3c56f-3060-4d4a-8dad-a411db392426 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 49.756s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.063490] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.370s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.064445] env[62952]: DEBUG nova.compute.manager [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 983.067573] env[62952]: DEBUG nova.network.neutron [-] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.069448] env[62952]: DEBUG oslo_concurrency.lockutils [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.270s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.070771] env[62952]: DEBUG oslo_concurrency.lockutils [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.074099] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.906s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.074099] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.074745] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.393s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.076520] env[62952]: INFO nova.compute.claims [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 983.090196] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bd6956d4-42ca-4499-9a2f-70d45a8eff00 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.100158] env[62952]: DEBUG nova.network.neutron [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Successfully updated port: ea5399c4-f18f-4c0b-82b1-062064db974b {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 983.104282] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13602b34-d753-4b0c-81a9-505f6a656fcc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.128271] env[62952]: INFO nova.scheduler.client.report [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleted allocations for instance e17b3d83-7cea-4a65-9457-7e57a6e7641a [ 983.133352] env[62952]: INFO nova.scheduler.client.report [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Deleted allocations for instance 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5 [ 983.152302] env[62952]: DEBUG nova.compute.manager [req-5552ea1d-760c-4416-98f5-9b40c916139c req-9d47f1b6-35f1-4056-9055-8c05a2328c87 service nova] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Detach interface failed, port_id=6bab534b-286c-4391-aeeb-fbb57655925a, reason: Instance dafa4926-3ab5-47f9-8ff2-4049e05c3113 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 983.181365] env[62952]: DEBUG nova.network.neutron [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Updating instance_info_cache with network_info: [{"id": "d93fe914-641a-4750-a765-181c7012b252", "address": "fa:16:3e:04:cd:36", "network": {"id": "827a8171-6759-4434-a6f1-437aad881c81", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1902486970-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091cdeb6da48492bb02f93822a45c9bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11b669be-fb26-4ef8-bdb6-c77ab9d06daf", "external-id": "nsx-vlan-transportzone-633", "segmentation_id": 633, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd93fe914-64", "ovs_interfaceid": "d93fe914-641a-4750-a765-181c7012b252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.203829] env[62952]: DEBUG nova.compute.manager [req-12ed548c-be05-455f-80c5-f90178ed3b1e req-fe145042-1aa6-4a15-93e5-7cbf636aa148 service nova] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Received event network-vif-plugged-ea5399c4-f18f-4c0b-82b1-062064db974b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 983.204075] env[62952]: DEBUG oslo_concurrency.lockutils [req-12ed548c-be05-455f-80c5-f90178ed3b1e req-fe145042-1aa6-4a15-93e5-7cbf636aa148 service nova] Acquiring lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.204294] env[62952]: DEBUG oslo_concurrency.lockutils [req-12ed548c-be05-455f-80c5-f90178ed3b1e req-fe145042-1aa6-4a15-93e5-7cbf636aa148 service nova] Lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.204497] env[62952]: DEBUG oslo_concurrency.lockutils [req-12ed548c-be05-455f-80c5-f90178ed3b1e req-fe145042-1aa6-4a15-93e5-7cbf636aa148 service nova] Lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.204666] env[62952]: DEBUG nova.compute.manager [req-12ed548c-be05-455f-80c5-f90178ed3b1e req-fe145042-1aa6-4a15-93e5-7cbf636aa148 service nova] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] No waiting events found dispatching network-vif-plugged-ea5399c4-f18f-4c0b-82b1-062064db974b {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 983.204804] env[62952]: WARNING nova.compute.manager [req-12ed548c-be05-455f-80c5-f90178ed3b1e req-fe145042-1aa6-4a15-93e5-7cbf636aa148 service nova] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Received unexpected event network-vif-plugged-ea5399c4-f18f-4c0b-82b1-062064db974b for instance with vm_state building and task_state spawning. [ 983.331664] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 983.331664] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-07a16262-f0d2-4c97-9c3a-92c6c11d6fab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.339539] env[62952]: DEBUG oslo_vmware.api [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 983.339539] env[62952]: value = "task-1263569" [ 983.339539] env[62952]: _type = "Task" [ 983.339539] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.349209] env[62952]: DEBUG oslo_vmware.api [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263569, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.410158] env[62952]: INFO nova.compute.manager [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Took 36.66 seconds to build instance. [ 983.569957] env[62952]: INFO nova.compute.manager [-] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Took 2.12 seconds to deallocate network for instance. [ 983.571315] env[62952]: DEBUG nova.compute.utils [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 983.574125] env[62952]: DEBUG nova.compute.manager [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 983.574336] env[62952]: DEBUG nova.network.neutron [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 983.600564] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "refresh_cache-75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.600854] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "refresh_cache-75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.600918] env[62952]: DEBUG nova.network.neutron [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 983.620963] env[62952]: DEBUG nova.policy [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b7e0ae49fe64e56bddf86570ed8270b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a6fddef5a2114acb9749b1bae171cbca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 983.638451] env[62952]: DEBUG oslo_concurrency.lockutils [None req-235d63fa-9bd9-4182-bb70-a77b3839997a tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "e17b3d83-7cea-4a65-9457-7e57a6e7641a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.792s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.654212] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d18c918-37aa-4cc3-aa04-e320e5e02f66 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.825s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.686332] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Releasing lock "refresh_cache-dbca4369-1ed3-493a-9847-9fa1e4293475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.852309] env[62952]: DEBUG oslo_vmware.api [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263569, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.912560] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c23cb2e7-956c-4037-b2e1-e90616c58066 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.166s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.078392] env[62952]: DEBUG nova.compute.manager [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 984.082978] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.110388] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquiring lock "14f8dbbe-2c4a-4948-bf15-106d5b1b0677" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.110663] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "14f8dbbe-2c4a-4948-bf15-106d5b1b0677" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.110879] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquiring lock "14f8dbbe-2c4a-4948-bf15-106d5b1b0677-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.111189] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "14f8dbbe-2c4a-4948-bf15-106d5b1b0677-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.111292] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "14f8dbbe-2c4a-4948-bf15-106d5b1b0677-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.113891] env[62952]: DEBUG nova.network.neutron [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Successfully created port: a6b1b43b-91ae-45fe-8e98-bd3a194bd05b {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 984.118315] env[62952]: INFO nova.compute.manager [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Terminating instance [ 984.120719] env[62952]: DEBUG nova.compute.manager [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 984.120772] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 984.122891] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49b72f6a-d680-41be-a152-2539178a9dd1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.131137] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 984.131280] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-17a66a1d-c03b-40c7-9435-95a481c9cef9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.140847] env[62952]: DEBUG oslo_vmware.api [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 984.140847] env[62952]: value = "task-1263570" [ 984.140847] env[62952]: _type = "Task" [ 984.140847] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.141693] env[62952]: DEBUG nova.network.neutron [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 984.153820] env[62952]: DEBUG oslo_vmware.api [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263570, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.216338] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 984.216636] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f92d7d90-2cd8-4fc5-9c1c-d984b7fa8bae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.226817] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 984.226817] env[62952]: value = "task-1263571" [ 984.226817] env[62952]: _type = "Task" [ 984.226817] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.240482] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263571, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.325507] env[62952]: DEBUG nova.network.neutron [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Updating instance_info_cache with network_info: [{"id": "ea5399c4-f18f-4c0b-82b1-062064db974b", "address": "fa:16:3e:bb:92:30", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea5399c4-f1", "ovs_interfaceid": "ea5399c4-f18f-4c0b-82b1-062064db974b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.335300] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6d3903-e9a6-4dc0-8625-043dab0b1ee0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.349321] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-073d7e71-eec6-45c4-938b-48689c1723db {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.358905] env[62952]: DEBUG oslo_vmware.api [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263569, 'name': CreateSnapshot_Task, 'duration_secs': 0.9035} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.359839] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 984.360953] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a15e5c5-4a1f-4aad-9c3a-315007952ef1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.409466] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87077128-60b6-43d3-8279-a234596b493d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.429477] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-907b795c-308d-4dc8-b63c-db8d2ea10487 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.447381] env[62952]: DEBUG nova.compute.provider_tree [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.656190] env[62952]: DEBUG oslo_vmware.api [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263570, 'name': PowerOffVM_Task, 'duration_secs': 0.249834} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.656482] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 984.656687] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 984.656962] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ef84be86-755b-4412-b701-e3d368e0bbd5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.726611] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 984.726842] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 984.727032] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Deleting the datastore file [datastore2] 14f8dbbe-2c4a-4948-bf15-106d5b1b0677 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 984.727438] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-15850f32-1916-4c9b-a81d-484aa11183fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.738173] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263571, 'name': PowerOffVM_Task, 'duration_secs': 0.332699} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.739452] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 984.739836] env[62952]: DEBUG oslo_vmware.api [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for the task: (returnval){ [ 984.739836] env[62952]: value = "task-1263573" [ 984.739836] env[62952]: _type = "Task" [ 984.739836] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.740536] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ea33a7-14ef-4b26-a8c7-025d41f08132 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.750799] env[62952]: DEBUG oslo_vmware.api [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263573, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.765516] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69214400-526d-4659-aa02-f4f95fd369fd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.800450] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 984.800790] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7169c821-c7bb-4377-8d2a-dc673c6a0189 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.809411] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 984.809411] env[62952]: value = "task-1263574" [ 984.809411] env[62952]: _type = "Task" [ 984.809411] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.818353] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263574, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.829060] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "refresh_cache-75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.829323] env[62952]: DEBUG nova.compute.manager [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Instance network_info: |[{"id": "ea5399c4-f18f-4c0b-82b1-062064db974b", "address": "fa:16:3e:bb:92:30", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea5399c4-f1", "ovs_interfaceid": "ea5399c4-f18f-4c0b-82b1-062064db974b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 984.829740] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:92:30', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5ad8894f-e240-4013-8272-4e79daea0751', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ea5399c4-f18f-4c0b-82b1-062064db974b', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 984.837695] env[62952]: DEBUG oslo.service.loopingcall [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.837938] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 984.838182] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-be697dff-d344-43a7-8f6a-55931c2f35d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.861665] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 984.861665] env[62952]: value = "task-1263575" [ 984.861665] env[62952]: _type = "Task" [ 984.861665] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.872390] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263575, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.915201] env[62952]: DEBUG oslo_concurrency.lockutils [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquiring lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.915496] env[62952]: DEBUG oslo_concurrency.lockutils [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.915687] env[62952]: DEBUG oslo_concurrency.lockutils [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquiring lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.915873] env[62952]: DEBUG oslo_concurrency.lockutils [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.916072] env[62952]: DEBUG oslo_concurrency.lockutils [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.918400] env[62952]: INFO nova.compute.manager [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Terminating instance [ 984.920307] env[62952]: DEBUG nova.compute.manager [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 984.920488] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 984.921322] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2e0ef3-cea2-45d9-ae7f-973d1ce4fb08 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.930677] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 984.930967] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d74fe346-619c-4bfc-a732-fb22d8ada8a7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.941301] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 984.942646] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3eae210a-e9a7-4ade-a1c9-cf3bdc87fbd7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.944354] env[62952]: DEBUG oslo_vmware.api [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 984.944354] env[62952]: value = "task-1263576" [ 984.944354] env[62952]: _type = "Task" [ 984.944354] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.950775] env[62952]: DEBUG nova.scheduler.client.report [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.954766] env[62952]: DEBUG oslo_vmware.api [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 984.954766] env[62952]: value = "task-1263577" [ 984.954766] env[62952]: _type = "Task" [ 984.954766] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.958914] env[62952]: DEBUG oslo_vmware.api [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263576, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.967341] env[62952]: DEBUG oslo_vmware.api [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263577, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.090009] env[62952]: DEBUG nova.compute.manager [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 985.115898] env[62952]: DEBUG nova.virt.hardware [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 985.116201] env[62952]: DEBUG nova.virt.hardware [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 985.116366] env[62952]: DEBUG nova.virt.hardware [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.116552] env[62952]: DEBUG nova.virt.hardware [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 985.116703] env[62952]: DEBUG nova.virt.hardware [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.116856] env[62952]: DEBUG nova.virt.hardware [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 985.117119] env[62952]: DEBUG nova.virt.hardware [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 985.117300] env[62952]: DEBUG nova.virt.hardware [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 985.117475] env[62952]: DEBUG nova.virt.hardware [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 985.117668] env[62952]: DEBUG nova.virt.hardware [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 985.117865] env[62952]: DEBUG nova.virt.hardware [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.118797] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6d3351-ec95-48a9-be13-dc0ad5b6f2ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.128059] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf267dd6-071a-4985-b175-e626aaa9469e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.236770] env[62952]: DEBUG nova.compute.manager [req-62141d0b-2e18-46b1-978f-ed0f15577c8c req-f42a8baf-e152-47c5-b2bf-4dc424a65030 service nova] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Received event network-changed-ea5399c4-f18f-4c0b-82b1-062064db974b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 985.237009] env[62952]: DEBUG nova.compute.manager [req-62141d0b-2e18-46b1-978f-ed0f15577c8c req-f42a8baf-e152-47c5-b2bf-4dc424a65030 service nova] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Refreshing instance network info cache due to event network-changed-ea5399c4-f18f-4c0b-82b1-062064db974b. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 985.237224] env[62952]: DEBUG oslo_concurrency.lockutils [req-62141d0b-2e18-46b1-978f-ed0f15577c8c req-f42a8baf-e152-47c5-b2bf-4dc424a65030 service nova] Acquiring lock "refresh_cache-75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.237374] env[62952]: DEBUG oslo_concurrency.lockutils [req-62141d0b-2e18-46b1-978f-ed0f15577c8c req-f42a8baf-e152-47c5-b2bf-4dc424a65030 service nova] Acquired lock "refresh_cache-75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.237542] env[62952]: DEBUG nova.network.neutron [req-62141d0b-2e18-46b1-978f-ed0f15577c8c req-f42a8baf-e152-47c5-b2bf-4dc424a65030 service nova] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Refreshing network info cache for port ea5399c4-f18f-4c0b-82b1-062064db974b {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 985.254136] env[62952]: DEBUG oslo_vmware.api [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Task: {'id': task-1263573, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.242983} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.254412] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 985.254602] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 985.254780] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 985.254953] env[62952]: INFO nova.compute.manager [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Took 1.13 seconds to destroy the instance on the hypervisor. [ 985.255218] env[62952]: DEBUG oslo.service.loopingcall [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.255416] env[62952]: DEBUG nova.compute.manager [-] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 985.255511] env[62952]: DEBUG nova.network.neutron [-] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 985.321243] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 985.321503] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 985.321756] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.321912] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.322111] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 985.322386] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3adc8909-de70-4960-aee3-1a536a1e4897 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.332833] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 985.333067] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 985.333846] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcd2c5ad-63d8-49e4-a365-8971811aa515 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.342895] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 985.342895] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a22ea2-fe32-89fa-9475-9d2fdec5330d" [ 985.342895] env[62952]: _type = "Task" [ 985.342895] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.353163] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a22ea2-fe32-89fa-9475-9d2fdec5330d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.372478] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263575, 'name': CreateVM_Task, 'duration_secs': 0.412724} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.372662] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 985.373360] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.373549] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.373875] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 985.374259] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c997932-e03e-437c-9500-a4287494d2b7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.379818] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 985.379818] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52610224-f16b-d64b-bb9a-a9549e317ae0" [ 985.379818] env[62952]: _type = "Task" [ 985.379818] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.389217] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52610224-f16b-d64b-bb9a-a9549e317ae0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.456899] env[62952]: DEBUG oslo_vmware.api [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263576, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.459805] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.460381] env[62952]: DEBUG nova.compute.manager [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 985.463127] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.289s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.463644] env[62952]: DEBUG nova.objects.instance [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lazy-loading 'resources' on Instance uuid b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.476245] env[62952]: DEBUG oslo_vmware.api [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263577, 'name': PowerOffVM_Task, 'duration_secs': 0.406529} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.476534] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 985.476725] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 985.476999] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b02e24d-086b-4577-ab5b-d9ee11069cec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.651937] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 985.652188] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 985.652378] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Deleting the datastore file [datastore1] fba2e955-8b5d-461c-997b-ef8f3ec0d906 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 985.652659] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ad24dde-be72-4b44-8531-8d0681af6975 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.660760] env[62952]: DEBUG oslo_vmware.api [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 985.660760] env[62952]: value = "task-1263579" [ 985.660760] env[62952]: _type = "Task" [ 985.660760] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.669822] env[62952]: DEBUG oslo_vmware.api [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263579, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.844816] env[62952]: DEBUG nova.compute.manager [req-42e287ce-87bc-4154-8747-a915d23251ef req-edfa560f-edea-4fb0-aba7-56d24d43d410 service nova] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Received event network-vif-plugged-a6b1b43b-91ae-45fe-8e98-bd3a194bd05b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 985.845095] env[62952]: DEBUG oslo_concurrency.lockutils [req-42e287ce-87bc-4154-8747-a915d23251ef req-edfa560f-edea-4fb0-aba7-56d24d43d410 service nova] Acquiring lock "7ba1587d-edaf-4dce-a224-6b2ac22cfecd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.845352] env[62952]: DEBUG oslo_concurrency.lockutils [req-42e287ce-87bc-4154-8747-a915d23251ef req-edfa560f-edea-4fb0-aba7-56d24d43d410 service nova] Lock "7ba1587d-edaf-4dce-a224-6b2ac22cfecd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.845609] env[62952]: DEBUG oslo_concurrency.lockutils [req-42e287ce-87bc-4154-8747-a915d23251ef req-edfa560f-edea-4fb0-aba7-56d24d43d410 service nova] Lock "7ba1587d-edaf-4dce-a224-6b2ac22cfecd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.845721] env[62952]: DEBUG nova.compute.manager [req-42e287ce-87bc-4154-8747-a915d23251ef req-edfa560f-edea-4fb0-aba7-56d24d43d410 service nova] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] No waiting events found dispatching network-vif-plugged-a6b1b43b-91ae-45fe-8e98-bd3a194bd05b {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 985.846025] env[62952]: WARNING nova.compute.manager [req-42e287ce-87bc-4154-8747-a915d23251ef req-edfa560f-edea-4fb0-aba7-56d24d43d410 service nova] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Received unexpected event network-vif-plugged-a6b1b43b-91ae-45fe-8e98-bd3a194bd05b for instance with vm_state building and task_state spawning. [ 985.861155] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a22ea2-fe32-89fa-9475-9d2fdec5330d, 'name': SearchDatastore_Task, 'duration_secs': 0.012698} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.862077] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed894573-dec5-41ee-8003-08010b17326c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.868646] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 985.868646] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5234ddf8-ff5a-7afb-ec8b-c0f05a0357a7" [ 985.868646] env[62952]: _type = "Task" [ 985.868646] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.878427] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5234ddf8-ff5a-7afb-ec8b-c0f05a0357a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.890884] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52610224-f16b-d64b-bb9a-a9549e317ae0, 'name': SearchDatastore_Task, 'duration_secs': 0.011082} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.893820] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.894122] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 985.894342] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.894514] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.894697] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 985.894988] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-875120a6-4820-4824-a9ce-244f9f6768d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.904321] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 985.904507] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 985.905240] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04b54573-5055-4cba-848c-f57ab691ccc8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.910892] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 985.910892] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52078072-c33f-4b13-e4fb-bf0f3c9d7445" [ 985.910892] env[62952]: _type = "Task" [ 985.910892] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.920050] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52078072-c33f-4b13-e4fb-bf0f3c9d7445, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.955858] env[62952]: DEBUG oslo_vmware.api [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263576, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.967365] env[62952]: DEBUG nova.network.neutron [req-62141d0b-2e18-46b1-978f-ed0f15577c8c req-f42a8baf-e152-47c5-b2bf-4dc424a65030 service nova] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Updated VIF entry in instance network info cache for port ea5399c4-f18f-4c0b-82b1-062064db974b. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 985.967752] env[62952]: DEBUG nova.network.neutron [req-62141d0b-2e18-46b1-978f-ed0f15577c8c req-f42a8baf-e152-47c5-b2bf-4dc424a65030 service nova] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Updating instance_info_cache with network_info: [{"id": "ea5399c4-f18f-4c0b-82b1-062064db974b", "address": "fa:16:3e:bb:92:30", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea5399c4-f1", "ovs_interfaceid": "ea5399c4-f18f-4c0b-82b1-062064db974b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.972793] env[62952]: DEBUG nova.compute.utils [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 985.974369] env[62952]: DEBUG nova.compute.manager [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 985.974541] env[62952]: DEBUG nova.network.neutron [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 986.003842] env[62952]: DEBUG nova.network.neutron [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Successfully updated port: a6b1b43b-91ae-45fe-8e98-bd3a194bd05b {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.014034] env[62952]: DEBUG nova.network.neutron [-] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.017933] env[62952]: DEBUG nova.policy [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7eadc05f193645e6b6e86e9cdbbbaa71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '076fb3c53420447caf047f8422435152', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 986.173114] env[62952]: DEBUG oslo_vmware.api [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263579, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.235903} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.173382] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 986.173572] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 986.173757] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 986.173938] env[62952]: INFO nova.compute.manager [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Took 1.25 seconds to destroy the instance on the hypervisor. [ 986.174189] env[62952]: DEBUG oslo.service.loopingcall [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.174392] env[62952]: DEBUG nova.compute.manager [-] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 986.174484] env[62952]: DEBUG nova.network.neutron [-] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 986.177229] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577a881e-7ba2-42e7-819f-bbbdff9853cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.187351] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2de1d951-5846-4ceb-bb76-716d0e62f00e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.220549] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f466fbe9-d2c6-45a6-a446-cadb59c8f33c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.228892] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f99354-52bb-4a4c-becf-3c72900545e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.242374] env[62952]: DEBUG nova.compute.provider_tree [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.277025] env[62952]: DEBUG nova.network.neutron [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Successfully created port: 7847ca5f-ad23-49b3-b3e9-6ef833b3c964 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 986.380938] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5234ddf8-ff5a-7afb-ec8b-c0f05a0357a7, 'name': SearchDatastore_Task, 'duration_secs': 0.010112} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.381990] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.382543] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] dbca4369-1ed3-493a-9847-9fa1e4293475/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk. {{(pid=62952) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 986.384533] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c05d7b8b-8ece-4af9-8f90-c9a274301c6f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.396556] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 986.396556] env[62952]: value = "task-1263580" [ 986.396556] env[62952]: _type = "Task" [ 986.396556] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.407982] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263580, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.422155] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52078072-c33f-4b13-e4fb-bf0f3c9d7445, 'name': SearchDatastore_Task, 'duration_secs': 0.011124} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.422520] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dca34d3d-86ad-402f-a791-c1d1cd2577c4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.429915] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 986.429915] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525d8f0c-ce3b-d4ea-91a6-ab037dd89bab" [ 986.429915] env[62952]: _type = "Task" [ 986.429915] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.442183] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525d8f0c-ce3b-d4ea-91a6-ab037dd89bab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.458562] env[62952]: DEBUG oslo_vmware.api [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263576, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.473098] env[62952]: DEBUG oslo_concurrency.lockutils [req-62141d0b-2e18-46b1-978f-ed0f15577c8c req-f42a8baf-e152-47c5-b2bf-4dc424a65030 service nova] Releasing lock "refresh_cache-75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.477886] env[62952]: DEBUG nova.compute.manager [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 986.508296] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "refresh_cache-7ba1587d-edaf-4dce-a224-6b2ac22cfecd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.508455] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquired lock "refresh_cache-7ba1587d-edaf-4dce-a224-6b2ac22cfecd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.508611] env[62952]: DEBUG nova.network.neutron [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 986.520895] env[62952]: INFO nova.compute.manager [-] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Took 1.27 seconds to deallocate network for instance. [ 986.746572] env[62952]: DEBUG nova.scheduler.client.report [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.910234] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263580, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.941936] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525d8f0c-ce3b-d4ea-91a6-ab037dd89bab, 'name': SearchDatastore_Task, 'duration_secs': 0.011235} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.942271] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.942545] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8/75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 986.942825] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8aecf1f3-6f20-45ea-ae2b-778192169a65 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.954325] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 986.954325] env[62952]: value = "task-1263581" [ 986.954325] env[62952]: _type = "Task" [ 986.954325] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.960188] env[62952]: DEBUG oslo_vmware.api [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263576, 'name': CloneVM_Task, 'duration_secs': 1.758227} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.961344] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Created linked-clone VM from snapshot [ 986.961690] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff9661a-3c25-4ac5-b6ee-9ee0d9900407 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.968122] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263581, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.974566] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Uploading image 47d5f45b-e188-47bb-bc3c-40e74aae2210 {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 986.989231] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 986.989557] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-dffe20f5-3702-415d-91b3-c1073ac8e4ae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.016047] env[62952]: DEBUG oslo_vmware.api [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 987.016047] env[62952]: value = "task-1263582" [ 987.016047] env[62952]: _type = "Task" [ 987.016047] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.029884] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.030269] env[62952]: DEBUG oslo_vmware.api [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263582, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.050790] env[62952]: DEBUG nova.network.neutron [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 987.217499] env[62952]: DEBUG nova.network.neutron [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Updating instance_info_cache with network_info: [{"id": "a6b1b43b-91ae-45fe-8e98-bd3a194bd05b", "address": "fa:16:3e:6b:a5:77", "network": {"id": "d84b900e-d793-4187-aaeb-bff9b040e607", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1370737473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a6fddef5a2114acb9749b1bae171cbca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6b1b43b-91", "ovs_interfaceid": "a6b1b43b-91ae-45fe-8e98-bd3a194bd05b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.250399] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.787s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.252783] env[62952]: DEBUG oslo_concurrency.lockutils [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.879s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.253068] env[62952]: DEBUG nova.objects.instance [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lazy-loading 'resources' on Instance uuid f4f8bee4-bf92-4321-afdc-16f7fb17dbae {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.268025] env[62952]: DEBUG nova.compute.manager [req-e474c3c3-ef3a-4be6-8e01-9f7225da1155 req-24cfa911-d2dd-43f3-9fd7-6e7b0d04c192 service nova] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Received event network-vif-deleted-f9a55a9d-b041-4065-b354-8a180d765835 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 987.278639] env[62952]: INFO nova.scheduler.client.report [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleted allocations for instance b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e [ 987.411193] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263580, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.644929} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.411545] env[62952]: INFO nova.virt.vmwareapi.ds_util [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] dbca4369-1ed3-493a-9847-9fa1e4293475/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk. [ 987.412670] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f4219e-d93a-4e1c-91b7-7ced09e2db74 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.444446] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] dbca4369-1ed3-493a-9847-9fa1e4293475/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 987.444894] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8003f3f-d69d-4e32-b21a-09fdc0219ed6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.469453] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263581, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.471525] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 987.471525] env[62952]: value = "task-1263583" [ 987.471525] env[62952]: _type = "Task" [ 987.471525] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.482059] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263583, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.488960] env[62952]: DEBUG nova.compute.manager [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 987.515289] env[62952]: DEBUG nova.virt.hardware [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.515549] env[62952]: DEBUG nova.virt.hardware [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.515711] env[62952]: DEBUG nova.virt.hardware [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.515910] env[62952]: DEBUG nova.virt.hardware [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.516102] env[62952]: DEBUG nova.virt.hardware [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.516264] env[62952]: DEBUG nova.virt.hardware [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.516475] env[62952]: DEBUG nova.virt.hardware [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.516637] env[62952]: DEBUG nova.virt.hardware [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.516805] env[62952]: DEBUG nova.virt.hardware [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.516970] env[62952]: DEBUG nova.virt.hardware [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.517189] env[62952]: DEBUG nova.virt.hardware [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.518190] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc794cf-d826-4f33-ac6b-b7026e72c4ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.536493] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd9f472b-6675-4d38-95c8-9565e29c2cd9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.540209] env[62952]: DEBUG oslo_vmware.api [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263582, 'name': Destroy_Task, 'duration_secs': 0.428452} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.540463] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Destroyed the VM [ 987.540693] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 987.541280] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-23bf60a1-93a7-4a14-871b-8592f925e254 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.554519] env[62952]: DEBUG oslo_vmware.api [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 987.554519] env[62952]: value = "task-1263584" [ 987.554519] env[62952]: _type = "Task" [ 987.554519] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.562106] env[62952]: DEBUG oslo_vmware.api [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263584, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.647457] env[62952]: DEBUG nova.network.neutron [-] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.720526] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Releasing lock "refresh_cache-7ba1587d-edaf-4dce-a224-6b2ac22cfecd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.720953] env[62952]: DEBUG nova.compute.manager [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Instance network_info: |[{"id": "a6b1b43b-91ae-45fe-8e98-bd3a194bd05b", "address": "fa:16:3e:6b:a5:77", "network": {"id": "d84b900e-d793-4187-aaeb-bff9b040e607", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1370737473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a6fddef5a2114acb9749b1bae171cbca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6b1b43b-91", "ovs_interfaceid": "a6b1b43b-91ae-45fe-8e98-bd3a194bd05b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 987.721409] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:a5:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc9714ff-7109-4ea1-9435-b2b3fbdb9e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a6b1b43b-91ae-45fe-8e98-bd3a194bd05b', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.729530] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Creating folder: Project (a6fddef5a2114acb9749b1bae171cbca). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 987.730134] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-59c59013-13df-49ee-8f1c-3b83c75295cb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.742836] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Created folder: Project (a6fddef5a2114acb9749b1bae171cbca) in parent group-v271811. [ 987.743271] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Creating folder: Instances. Parent ref: group-v271932. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 987.743862] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce694e6c-ef9e-49d3-b6bf-dc867f715d2d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.755235] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Created folder: Instances in parent group-v271932. [ 987.755588] env[62952]: DEBUG oslo.service.loopingcall [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.758421] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 987.758966] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bb55f148-fa0e-4d63-a468-a580bb39dd24 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.782956] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.782956] env[62952]: value = "task-1263587" [ 987.782956] env[62952]: _type = "Task" [ 987.782956] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.789829] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1961e40e-8f10-439f-856d-d25b3e21487b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.711s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.806113] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263587, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.881712] env[62952]: DEBUG nova.compute.manager [req-626d6083-cf54-4d3b-bd2d-d5aa318f5170 req-7c7d52a6-d6ba-419a-bc0d-98066fc92939 service nova] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Received event network-changed-a6b1b43b-91ae-45fe-8e98-bd3a194bd05b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 987.881970] env[62952]: DEBUG nova.compute.manager [req-626d6083-cf54-4d3b-bd2d-d5aa318f5170 req-7c7d52a6-d6ba-419a-bc0d-98066fc92939 service nova] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Refreshing instance network info cache due to event network-changed-a6b1b43b-91ae-45fe-8e98-bd3a194bd05b. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 987.882412] env[62952]: DEBUG oslo_concurrency.lockutils [req-626d6083-cf54-4d3b-bd2d-d5aa318f5170 req-7c7d52a6-d6ba-419a-bc0d-98066fc92939 service nova] Acquiring lock "refresh_cache-7ba1587d-edaf-4dce-a224-6b2ac22cfecd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.882557] env[62952]: DEBUG oslo_concurrency.lockutils [req-626d6083-cf54-4d3b-bd2d-d5aa318f5170 req-7c7d52a6-d6ba-419a-bc0d-98066fc92939 service nova] Acquired lock "refresh_cache-7ba1587d-edaf-4dce-a224-6b2ac22cfecd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.882741] env[62952]: DEBUG nova.network.neutron [req-626d6083-cf54-4d3b-bd2d-d5aa318f5170 req-7c7d52a6-d6ba-419a-bc0d-98066fc92939 service nova] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Refreshing network info cache for port a6b1b43b-91ae-45fe-8e98-bd3a194bd05b {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 987.970578] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263581, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.625496} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.970578] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8/75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 987.970578] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 987.970782] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-deec01fb-7dc1-4959-91cc-a8a2b4b9c651 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.986470] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263583, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.987581] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 987.987581] env[62952]: value = "task-1263588" [ 987.987581] env[62952]: _type = "Task" [ 987.987581] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.997483] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263588, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.028219] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1daca6-05a7-4566-93a1-e3f465d2431d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.038678] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c29bfc1-09f4-4865-b429-1e524943ad01 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.074805] env[62952]: DEBUG nova.network.neutron [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Successfully updated port: 7847ca5f-ad23-49b3-b3e9-6ef833b3c964 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 988.080019] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1db458e3-420d-47eb-aa39-28d403d1e032 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.093632] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bde8f8b-5645-430c-846b-bb31b5711d0c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.099148] env[62952]: DEBUG oslo_vmware.api [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263584, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.110617] env[62952]: DEBUG nova.compute.provider_tree [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.150193] env[62952]: INFO nova.compute.manager [-] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Took 1.98 seconds to deallocate network for instance. [ 988.293578] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263587, 'name': CreateVM_Task, 'duration_secs': 0.419659} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.293760] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 988.294495] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.294667] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.294998] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 988.295299] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-efee9ee6-e447-48e6-a408-41489b75bb75 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.300536] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 988.300536] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d79c28-961b-0ef8-c9a7-d87297323e90" [ 988.300536] env[62952]: _type = "Task" [ 988.300536] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.308645] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d79c28-961b-0ef8-c9a7-d87297323e90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.488419] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263583, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.497392] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263588, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.09341} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.497392] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 988.497887] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd66a6f-92f7-4bdf-a4e8-8d7533d49db2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.524095] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8/75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 988.528099] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ef2f941-302c-4053-ad2c-0aa166984f15 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.550100] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 988.550100] env[62952]: value = "task-1263589" [ 988.550100] env[62952]: _type = "Task" [ 988.550100] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.562721] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263589, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.584630] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "refresh_cache-6f0b04f5-811f-4c53-808b-6d9d22100a86" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.584814] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "refresh_cache-6f0b04f5-811f-4c53-808b-6d9d22100a86" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.584974] env[62952]: DEBUG nova.network.neutron [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 988.586606] env[62952]: DEBUG oslo_vmware.api [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263584, 'name': RemoveSnapshot_Task, 'duration_secs': 0.9105} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.587408] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 988.614329] env[62952]: DEBUG nova.scheduler.client.report [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 988.645959] env[62952]: DEBUG nova.network.neutron [req-626d6083-cf54-4d3b-bd2d-d5aa318f5170 req-7c7d52a6-d6ba-419a-bc0d-98066fc92939 service nova] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Updated VIF entry in instance network info cache for port a6b1b43b-91ae-45fe-8e98-bd3a194bd05b. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 988.646348] env[62952]: DEBUG nova.network.neutron [req-626d6083-cf54-4d3b-bd2d-d5aa318f5170 req-7c7d52a6-d6ba-419a-bc0d-98066fc92939 service nova] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Updating instance_info_cache with network_info: [{"id": "a6b1b43b-91ae-45fe-8e98-bd3a194bd05b", "address": "fa:16:3e:6b:a5:77", "network": {"id": "d84b900e-d793-4187-aaeb-bff9b040e607", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1370737473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a6fddef5a2114acb9749b1bae171cbca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6b1b43b-91", "ovs_interfaceid": "a6b1b43b-91ae-45fe-8e98-bd3a194bd05b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.657909] env[62952]: DEBUG oslo_concurrency.lockutils [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.815048] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d79c28-961b-0ef8-c9a7-d87297323e90, 'name': SearchDatastore_Task, 'duration_secs': 0.048865} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.815517] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.815852] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 988.816201] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.816444] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.816693] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 988.816984] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b396fca5-4137-45db-8a6d-b237f5396429 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.827572] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 988.828106] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 988.829120] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb459178-953e-4e82-b354-ea1bd2078047 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.836368] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 988.836368] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52710d2d-b384-825e-abaf-30ff63b6a2ba" [ 988.836368] env[62952]: _type = "Task" [ 988.836368] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.846851] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52710d2d-b384-825e-abaf-30ff63b6a2ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.985695] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263583, 'name': ReconfigVM_Task, 'duration_secs': 1.045083} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.986095] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Reconfigured VM instance instance-0000004d to attach disk [datastore1] dbca4369-1ed3-493a-9847-9fa1e4293475/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 988.987016] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73bb49d1-bf0f-40cb-abd8-e6b6f54e71cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.017105] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe2fb1f3-4882-4462-82b7-57a6481dbd3c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.032874] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 989.032874] env[62952]: value = "task-1263590" [ 989.032874] env[62952]: _type = "Task" [ 989.032874] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.042432] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263590, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.060316] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263589, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.095021] env[62952]: WARNING nova.compute.manager [None req-cbe2622c-9475-466d-9604-83edb5ce141c tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Image not found during snapshot: nova.exception.ImageNotFound: Image 47d5f45b-e188-47bb-bc3c-40e74aae2210 could not be found. [ 989.122077] env[62952]: DEBUG oslo_concurrency.lockutils [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.869s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.126622] env[62952]: DEBUG nova.network.neutron [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 989.132643] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.050s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.132921] env[62952]: DEBUG nova.objects.instance [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lazy-loading 'resources' on Instance uuid dafa4926-3ab5-47f9-8ff2-4049e05c3113 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.150376] env[62952]: DEBUG oslo_concurrency.lockutils [req-626d6083-cf54-4d3b-bd2d-d5aa318f5170 req-7c7d52a6-d6ba-419a-bc0d-98066fc92939 service nova] Releasing lock "refresh_cache-7ba1587d-edaf-4dce-a224-6b2ac22cfecd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.150376] env[62952]: DEBUG nova.compute.manager [req-626d6083-cf54-4d3b-bd2d-d5aa318f5170 req-7c7d52a6-d6ba-419a-bc0d-98066fc92939 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Received event network-vif-deleted-34b7738d-65b4-4571-bef9-bf3c5ff297ef {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 989.150376] env[62952]: DEBUG nova.compute.manager [req-626d6083-cf54-4d3b-bd2d-d5aa318f5170 req-7c7d52a6-d6ba-419a-bc0d-98066fc92939 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Received event network-vif-deleted-b0b3ad2c-5b56-4239-856d-663404db8324 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 989.150376] env[62952]: DEBUG nova.compute.manager [req-626d6083-cf54-4d3b-bd2d-d5aa318f5170 req-7c7d52a6-d6ba-419a-bc0d-98066fc92939 service nova] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Received event network-vif-deleted-6a5ee171-10c8-43aa-8b0b-12f2d7068e6e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 989.155658] env[62952]: INFO nova.scheduler.client.report [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Deleted allocations for instance f4f8bee4-bf92-4321-afdc-16f7fb17dbae [ 989.297369] env[62952]: DEBUG nova.compute.manager [req-86042de0-56f1-4a40-b394-c0eadbf56268 req-1256c314-f35e-48c5-9dc7-2a387f039d05 service nova] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Received event network-vif-plugged-7847ca5f-ad23-49b3-b3e9-6ef833b3c964 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 989.297598] env[62952]: DEBUG oslo_concurrency.lockutils [req-86042de0-56f1-4a40-b394-c0eadbf56268 req-1256c314-f35e-48c5-9dc7-2a387f039d05 service nova] Acquiring lock "6f0b04f5-811f-4c53-808b-6d9d22100a86-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.297926] env[62952]: DEBUG oslo_concurrency.lockutils [req-86042de0-56f1-4a40-b394-c0eadbf56268 req-1256c314-f35e-48c5-9dc7-2a387f039d05 service nova] Lock "6f0b04f5-811f-4c53-808b-6d9d22100a86-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.298379] env[62952]: DEBUG oslo_concurrency.lockutils [req-86042de0-56f1-4a40-b394-c0eadbf56268 req-1256c314-f35e-48c5-9dc7-2a387f039d05 service nova] Lock "6f0b04f5-811f-4c53-808b-6d9d22100a86-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.298579] env[62952]: DEBUG nova.compute.manager [req-86042de0-56f1-4a40-b394-c0eadbf56268 req-1256c314-f35e-48c5-9dc7-2a387f039d05 service nova] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] No waiting events found dispatching network-vif-plugged-7847ca5f-ad23-49b3-b3e9-6ef833b3c964 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 989.298751] env[62952]: WARNING nova.compute.manager [req-86042de0-56f1-4a40-b394-c0eadbf56268 req-1256c314-f35e-48c5-9dc7-2a387f039d05 service nova] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Received unexpected event network-vif-plugged-7847ca5f-ad23-49b3-b3e9-6ef833b3c964 for instance with vm_state building and task_state spawning. [ 989.298980] env[62952]: DEBUG nova.compute.manager [req-86042de0-56f1-4a40-b394-c0eadbf56268 req-1256c314-f35e-48c5-9dc7-2a387f039d05 service nova] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Received event network-changed-7847ca5f-ad23-49b3-b3e9-6ef833b3c964 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 989.299182] env[62952]: DEBUG nova.compute.manager [req-86042de0-56f1-4a40-b394-c0eadbf56268 req-1256c314-f35e-48c5-9dc7-2a387f039d05 service nova] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Refreshing instance network info cache due to event network-changed-7847ca5f-ad23-49b3-b3e9-6ef833b3c964. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 989.299358] env[62952]: DEBUG oslo_concurrency.lockutils [req-86042de0-56f1-4a40-b394-c0eadbf56268 req-1256c314-f35e-48c5-9dc7-2a387f039d05 service nova] Acquiring lock "refresh_cache-6f0b04f5-811f-4c53-808b-6d9d22100a86" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.329986] env[62952]: DEBUG nova.network.neutron [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Updating instance_info_cache with network_info: [{"id": "7847ca5f-ad23-49b3-b3e9-6ef833b3c964", "address": "fa:16:3e:71:6b:80", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7847ca5f-ad", "ovs_interfaceid": "7847ca5f-ad23-49b3-b3e9-6ef833b3c964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.349635] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52710d2d-b384-825e-abaf-30ff63b6a2ba, 'name': SearchDatastore_Task, 'duration_secs': 0.013406} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.350473] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-609416ed-fa11-48ea-89ff-0d5951e71965 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.357646] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 989.357646] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d8950f-ac9d-1fe3-3ced-85af549d1f48" [ 989.357646] env[62952]: _type = "Task" [ 989.357646] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.367985] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d8950f-ac9d-1fe3-3ced-85af549d1f48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.543971] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263590, 'name': ReconfigVM_Task, 'duration_secs': 0.366383} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.544349] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 989.544584] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7688a2cf-0680-40d4-ba85-00924cd01032 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.556718] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 989.556718] env[62952]: value = "task-1263591" [ 989.556718] env[62952]: _type = "Task" [ 989.556718] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.564337] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263589, 'name': ReconfigVM_Task, 'duration_secs': 0.567542} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.565268] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8/75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.565875] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a9eaf42d-70c0-4ffd-b8f4-c47c7a4f1534 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.571139] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263591, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.577541] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 989.577541] env[62952]: value = "task-1263592" [ 989.577541] env[62952]: _type = "Task" [ 989.577541] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.587028] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263592, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.643522] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "063d1d92-0526-48e0-b33c-f51a79e7c821" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.643732] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "063d1d92-0526-48e0-b33c-f51a79e7c821" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.671435] env[62952]: DEBUG oslo_concurrency.lockutils [None req-23c8c3ae-1ae5-43b0-a562-4244b6ae98f9 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "f4f8bee4-bf92-4321-afdc-16f7fb17dbae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.103s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.834531] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "refresh_cache-6f0b04f5-811f-4c53-808b-6d9d22100a86" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.834735] env[62952]: DEBUG nova.compute.manager [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Instance network_info: |[{"id": "7847ca5f-ad23-49b3-b3e9-6ef833b3c964", "address": "fa:16:3e:71:6b:80", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7847ca5f-ad", "ovs_interfaceid": "7847ca5f-ad23-49b3-b3e9-6ef833b3c964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 989.838567] env[62952]: DEBUG oslo_concurrency.lockutils [req-86042de0-56f1-4a40-b394-c0eadbf56268 req-1256c314-f35e-48c5-9dc7-2a387f039d05 service nova] Acquired lock "refresh_cache-6f0b04f5-811f-4c53-808b-6d9d22100a86" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.838764] env[62952]: DEBUG nova.network.neutron [req-86042de0-56f1-4a40-b394-c0eadbf56268 req-1256c314-f35e-48c5-9dc7-2a387f039d05 service nova] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Refreshing network info cache for port 7847ca5f-ad23-49b3-b3e9-6ef833b3c964 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 989.840168] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:6b:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '365ac5b1-6d83-4dfe-887f-60574d7f6124', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7847ca5f-ad23-49b3-b3e9-6ef833b3c964', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 989.848508] env[62952]: DEBUG oslo.service.loopingcall [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.849994] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 989.850375] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eb4df299-1d8c-4642-9e6e-ff90f9199a66 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.880646] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d8950f-ac9d-1fe3-3ced-85af549d1f48, 'name': SearchDatastore_Task, 'duration_secs': 0.014681} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.884123] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.884430] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 7ba1587d-edaf-4dce-a224-6b2ac22cfecd/7ba1587d-edaf-4dce-a224-6b2ac22cfecd.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 989.885286] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-569fc374-0c09-4fdd-95b8-ab7a587bd07b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.892914] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 989.892914] env[62952]: value = "task-1263593" [ 989.892914] env[62952]: _type = "Task" [ 989.892914] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.893343] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-34d98d41-84cd-4b9e-b252-564ff2e831cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.905059] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102026a0-180e-4b73-a95c-3f25536411a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.914750] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263593, 'name': CreateVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.915243] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 989.915243] env[62952]: value = "task-1263594" [ 989.915243] env[62952]: _type = "Task" [ 989.915243] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.950842] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45879ac4-8e60-42d8-acd9-d41c6c8cee44 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.953754] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263594, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.961420] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b829cc60-1813-463a-8d75-12bd6d45bc6f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.977194] env[62952]: DEBUG nova.compute.provider_tree [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.069880] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263591, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.092711] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263592, 'name': Rename_Task, 'duration_secs': 0.184005} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.093272] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 990.093681] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-041cef4e-c580-44f3-8a1a-f1aa5f08b855 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.102323] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 990.102323] env[62952]: value = "task-1263595" [ 990.102323] env[62952]: _type = "Task" [ 990.102323] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.113471] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263595, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.147065] env[62952]: DEBUG nova.compute.manager [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 990.396186] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "c7d08234-fcef-4cdb-939f-05ea9e8b08ea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.396468] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "c7d08234-fcef-4cdb-939f-05ea9e8b08ea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.396702] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "c7d08234-fcef-4cdb-939f-05ea9e8b08ea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.396897] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "c7d08234-fcef-4cdb-939f-05ea9e8b08ea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.397085] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "c7d08234-fcef-4cdb-939f-05ea9e8b08ea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.402770] env[62952]: INFO nova.compute.manager [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Terminating instance [ 990.405291] env[62952]: DEBUG nova.compute.manager [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 990.405595] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 990.406421] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cc9877f-0000-4144-9259-21ddafb6d4b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.414161] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263593, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.423123] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 990.425228] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00e7f230-e6f4-47c3-8ee1-bd91e3f53020 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.431749] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263594, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.433520] env[62952]: DEBUG oslo_vmware.api [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 990.433520] env[62952]: value = "task-1263596" [ 990.433520] env[62952]: _type = "Task" [ 990.433520] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.444960] env[62952]: DEBUG oslo_vmware.api [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263596, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.484019] env[62952]: DEBUG nova.scheduler.client.report [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.574690] env[62952]: DEBUG oslo_vmware.api [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263591, 'name': PowerOnVM_Task, 'duration_secs': 0.543379} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.575427] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 990.580065] env[62952]: DEBUG nova.compute.manager [None req-8c94168e-a0f7-49ac-8099-c1508d27fc30 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 990.581480] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e498b24f-6dfc-4762-a02c-7ed55630ce53 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.613358] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263595, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.633447] env[62952]: DEBUG nova.network.neutron [req-86042de0-56f1-4a40-b394-c0eadbf56268 req-1256c314-f35e-48c5-9dc7-2a387f039d05 service nova] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Updated VIF entry in instance network info cache for port 7847ca5f-ad23-49b3-b3e9-6ef833b3c964. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 990.633853] env[62952]: DEBUG nova.network.neutron [req-86042de0-56f1-4a40-b394-c0eadbf56268 req-1256c314-f35e-48c5-9dc7-2a387f039d05 service nova] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Updating instance_info_cache with network_info: [{"id": "7847ca5f-ad23-49b3-b3e9-6ef833b3c964", "address": "fa:16:3e:71:6b:80", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7847ca5f-ad", "ovs_interfaceid": "7847ca5f-ad23-49b3-b3e9-6ef833b3c964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.669010] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.910039] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263593, 'name': CreateVM_Task, 'duration_secs': 0.689249} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.910277] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 990.911021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.911204] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.911534] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 990.911814] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a0dd525-ce56-4524-a2b0-c74857ee2423 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.917478] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 990.917478] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c7d519-394e-b109-37e3-a1cfc5b77c6a" [ 990.917478] env[62952]: _type = "Task" [ 990.917478] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.931534] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263594, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542805} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.931784] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c7d519-394e-b109-37e3-a1cfc5b77c6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.932034] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 7ba1587d-edaf-4dce-a224-6b2ac22cfecd/7ba1587d-edaf-4dce-a224-6b2ac22cfecd.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 990.932274] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 990.932525] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0dd5f5c7-0a36-4e23-b45b-c458a6addafa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.942431] env[62952]: DEBUG oslo_vmware.api [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263596, 'name': PowerOffVM_Task, 'duration_secs': 0.353186} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.943675] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 990.943861] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 990.944539] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 990.944539] env[62952]: value = "task-1263597" [ 990.944539] env[62952]: _type = "Task" [ 990.944539] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.944749] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c2482d80-4840-41aa-8c51-9bfd78d535b3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.955657] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263597, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.991334] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.858s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.994537] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.965s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.995086] env[62952]: DEBUG nova.objects.instance [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lazy-loading 'resources' on Instance uuid 14f8dbbe-2c4a-4948-bf15-106d5b1b0677 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.016872] env[62952]: INFO nova.scheduler.client.report [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Deleted allocations for instance dafa4926-3ab5-47f9-8ff2-4049e05c3113 [ 991.020148] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 991.020148] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 991.020148] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Deleting the datastore file [datastore1] c7d08234-fcef-4cdb-939f-05ea9e8b08ea {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 991.023162] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9f02197b-1274-4247-b109-3f30aa074e60 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.033754] env[62952]: DEBUG oslo_vmware.api [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for the task: (returnval){ [ 991.033754] env[62952]: value = "task-1263599" [ 991.033754] env[62952]: _type = "Task" [ 991.033754] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.044419] env[62952]: DEBUG oslo_vmware.api [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263599, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.112966] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263595, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.136611] env[62952]: DEBUG oslo_concurrency.lockutils [req-86042de0-56f1-4a40-b394-c0eadbf56268 req-1256c314-f35e-48c5-9dc7-2a387f039d05 service nova] Releasing lock "refresh_cache-6f0b04f5-811f-4c53-808b-6d9d22100a86" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.436995] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c7d519-394e-b109-37e3-a1cfc5b77c6a, 'name': SearchDatastore_Task, 'duration_secs': 0.016594} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.437474] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.437819] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 991.438192] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.438428] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.438702] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 991.439085] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a2cd477f-ac70-40cf-a974-0c8ddf159938 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.453865] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 991.453865] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 991.455401] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9338b7ca-f901-480a-b7a1-3f21f8793f5a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.462852] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263597, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082395} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.463529] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 991.464734] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e5b81e1-08d6-44e8-a705-4ab0e0cc843c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.469798] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 991.469798] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524b65f9-be9f-e1e2-db1e-2a6d42915843" [ 991.469798] env[62952]: _type = "Task" [ 991.469798] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.503016] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 7ba1587d-edaf-4dce-a224-6b2ac22cfecd/7ba1587d-edaf-4dce-a224-6b2ac22cfecd.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.507484] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbaf9a48-9584-4918-be8e-4065fbf6b950 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.539410] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524b65f9-be9f-e1e2-db1e-2a6d42915843, 'name': SearchDatastore_Task, 'duration_secs': 0.034148} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.540056] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5003da62-52a7-46ba-bbde-eacc7d4cb670 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "dafa4926-3ab5-47f9-8ff2-4049e05c3113" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.222s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.548733] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03196feb-4815-403c-9b94-5a5eb91ed00b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.552042] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 991.552042] env[62952]: value = "task-1263600" [ 991.552042] env[62952]: _type = "Task" [ 991.552042] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.568538] env[62952]: DEBUG oslo_vmware.api [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Task: {'id': task-1263599, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.503384} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.568974] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 991.568974] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5291c53a-a15f-11d3-5cf0-7f13e0382daf" [ 991.568974] env[62952]: _type = "Task" [ 991.568974] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.569778] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 991.570106] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 991.570396] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 991.570668] env[62952]: INFO nova.compute.manager [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Took 1.17 seconds to destroy the instance on the hypervisor. [ 991.571015] env[62952]: DEBUG oslo.service.loopingcall [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.574756] env[62952]: DEBUG nova.compute.manager [-] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 991.575298] env[62952]: DEBUG nova.network.neutron [-] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 991.587151] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263600, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.595993] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5291c53a-a15f-11d3-5cf0-7f13e0382daf, 'name': SearchDatastore_Task, 'duration_secs': 0.014816} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.599775] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.600198] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 6f0b04f5-811f-4c53-808b-6d9d22100a86/6f0b04f5-811f-4c53-808b-6d9d22100a86.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 991.600990] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db10ae5e-e58a-4504-8535-3b94a1c274d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.619579] env[62952]: DEBUG oslo_vmware.api [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263595, 'name': PowerOnVM_Task, 'duration_secs': 1.082292} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.620015] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 991.620015] env[62952]: value = "task-1263601" [ 991.620015] env[62952]: _type = "Task" [ 991.620015] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.620382] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 991.620673] env[62952]: INFO nova.compute.manager [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Took 8.90 seconds to spawn the instance on the hypervisor. [ 991.620944] env[62952]: DEBUG nova.compute.manager [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 991.622220] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ecf6e8-5d2d-4b14-b341-4a5c9448de13 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.642743] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263601, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.772223] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a13da953-3756-4f87-987f-f261529d602e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.781356] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c19add-0ce2-4dcd-a2a6-a3439342200a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.814490] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89b97bb-3010-4674-a916-0c1780106c33 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.822967] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7c48417-5ecb-4bdc-aec9-87607fb3e856 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.839615] env[62952]: DEBUG nova.compute.provider_tree [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.071809] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263600, 'name': ReconfigVM_Task, 'duration_secs': 0.301982} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.072135] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 7ba1587d-edaf-4dce-a224-6b2ac22cfecd/7ba1587d-edaf-4dce-a224-6b2ac22cfecd.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 992.072830] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eab78ecc-7d0f-495e-8c67-01f26a1e8db4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.082561] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 992.082561] env[62952]: value = "task-1263602" [ 992.082561] env[62952]: _type = "Task" [ 992.082561] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.091916] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263602, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.132931] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263601, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500844} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.132931] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 6f0b04f5-811f-4c53-808b-6d9d22100a86/6f0b04f5-811f-4c53-808b-6d9d22100a86.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 992.133144] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 992.134033] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a37477b5-7eed-486c-a5a6-1ccb65b9631c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.144279] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 992.144279] env[62952]: value = "task-1263603" [ 992.144279] env[62952]: _type = "Task" [ 992.144279] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.158622] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263603, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.163979] env[62952]: INFO nova.compute.manager [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Took 31.39 seconds to build instance. [ 992.202880] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.203249] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.203529] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.203744] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.203931] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.207998] env[62952]: INFO nova.compute.manager [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Terminating instance [ 992.209826] env[62952]: DEBUG nova.compute.manager [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 992.210065] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 992.211546] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac816825-534a-47b0-9a07-11b7589c74ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.221553] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 992.221866] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d58690fb-70cf-4a0e-8deb-e22a7b464562 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.223801] env[62952]: INFO nova.compute.manager [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Unrescuing [ 992.224407] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "refresh_cache-dbca4369-1ed3-493a-9847-9fa1e4293475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.224407] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired lock "refresh_cache-dbca4369-1ed3-493a-9847-9fa1e4293475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.224560] env[62952]: DEBUG nova.network.neutron [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 992.231852] env[62952]: DEBUG oslo_vmware.api [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 992.231852] env[62952]: value = "task-1263604" [ 992.231852] env[62952]: _type = "Task" [ 992.231852] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.244021] env[62952]: DEBUG oslo_vmware.api [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263604, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.295096] env[62952]: DEBUG nova.compute.manager [req-1ede3506-4c9f-4a57-825e-1b0482216875 req-4dc2e874-9699-4d35-afd6-f5c06c56fff5 service nova] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Received event network-vif-deleted-89ebaa6b-2763-41e0-a0d7-954adf9ac031 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 992.295359] env[62952]: INFO nova.compute.manager [req-1ede3506-4c9f-4a57-825e-1b0482216875 req-4dc2e874-9699-4d35-afd6-f5c06c56fff5 service nova] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Neutron deleted interface 89ebaa6b-2763-41e0-a0d7-954adf9ac031; detaching it from the instance and deleting it from the info cache [ 992.295534] env[62952]: DEBUG nova.network.neutron [req-1ede3506-4c9f-4a57-825e-1b0482216875 req-4dc2e874-9699-4d35-afd6-f5c06c56fff5 service nova] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.343605] env[62952]: DEBUG nova.scheduler.client.report [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.403785] env[62952]: DEBUG nova.network.neutron [-] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.593571] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263602, 'name': Rename_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.654470] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263603, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070704} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.654871] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 992.655739] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c28be652-3268-42da-837b-1aa84057e2c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.671893] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af50c624-3ff7-475e-a8b2-091f672d44b8 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.907s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.683367] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] 6f0b04f5-811f-4c53-808b-6d9d22100a86/6f0b04f5-811f-4c53-808b-6d9d22100a86.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 992.684120] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71169749-ec1c-498b-89be-133bb334034a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.710874] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 992.710874] env[62952]: value = "task-1263605" [ 992.710874] env[62952]: _type = "Task" [ 992.710874] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.721396] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263605, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.742441] env[62952]: DEBUG oslo_vmware.api [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263604, 'name': PowerOffVM_Task, 'duration_secs': 0.215636} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.742704] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 992.742887] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 992.743176] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a1d52d2-6929-4e1e-b919-bb75e3a6d401 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.798375] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c26e9a0-a666-4018-a4bc-ca93f6f88671 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.813282] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29bfb2e5-4a6a-47e4-9863-7013fce321da {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.825747] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 992.827558] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 992.828039] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Deleting the datastore file [datastore2] 0533ee81-b69d-4f4e-9832-78c5ab4512b1 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 992.831407] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1b764d90-c5ed-4fb7-ac5e-3b85e8ab7f79 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.839318] env[62952]: DEBUG oslo_vmware.api [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for the task: (returnval){ [ 992.839318] env[62952]: value = "task-1263607" [ 992.839318] env[62952]: _type = "Task" [ 992.839318] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.856366] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.862s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.863355] env[62952]: DEBUG nova.compute.manager [req-1ede3506-4c9f-4a57-825e-1b0482216875 req-4dc2e874-9699-4d35-afd6-f5c06c56fff5 service nova] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Detach interface failed, port_id=89ebaa6b-2763-41e0-a0d7-954adf9ac031, reason: Instance c7d08234-fcef-4cdb-939f-05ea9e8b08ea could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 992.869557] env[62952]: DEBUG oslo_concurrency.lockutils [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.212s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.869891] env[62952]: DEBUG nova.objects.instance [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lazy-loading 'resources' on Instance uuid fba2e955-8b5d-461c-997b-ef8f3ec0d906 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.871452] env[62952]: DEBUG oslo_vmware.api [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263607, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.888637] env[62952]: INFO nova.scheduler.client.report [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Deleted allocations for instance 14f8dbbe-2c4a-4948-bf15-106d5b1b0677 [ 992.906112] env[62952]: INFO nova.compute.manager [-] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Took 1.33 seconds to deallocate network for instance. [ 993.021285] env[62952]: DEBUG nova.network.neutron [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Updating instance_info_cache with network_info: [{"id": "d93fe914-641a-4750-a765-181c7012b252", "address": "fa:16:3e:04:cd:36", "network": {"id": "827a8171-6759-4434-a6f1-437aad881c81", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1902486970-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091cdeb6da48492bb02f93822a45c9bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11b669be-fb26-4ef8-bdb6-c77ab9d06daf", "external-id": "nsx-vlan-transportzone-633", "segmentation_id": 633, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd93fe914-64", "ovs_interfaceid": "d93fe914-641a-4750-a765-181c7012b252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.095389] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263602, 'name': Rename_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.225631] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263605, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.278703] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.278962] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.350511] env[62952]: DEBUG oslo_vmware.api [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263607, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.399354] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9d14a669-f4b5-4ed0-87a7-97cb0febcc03 tempest-ListImageFiltersTestJSON-1809764396 tempest-ListImageFiltersTestJSON-1809764396-project-member] Lock "14f8dbbe-2c4a-4948-bf15-106d5b1b0677" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.288s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.414501] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.524635] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Releasing lock "refresh_cache-dbca4369-1ed3-493a-9847-9fa1e4293475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.525381] env[62952]: DEBUG nova.objects.instance [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lazy-loading 'flavor' on Instance uuid dbca4369-1ed3-493a-9847-9fa1e4293475 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.554516] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4569d582-db2f-4375-8933-6a4effb41a91 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.563261] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ea4ff00-8265-444e-bcf1-6385dadd9487 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.604467] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dfa6d5a-408b-4cc8-8584-c3018c6e2c97 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.610545] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263602, 'name': Rename_Task, 'duration_secs': 1.1592} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.612629] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 993.612940] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-155593c7-376d-4b72-a90b-004a93e58b52 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.615348] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a0fb795-c2c0-48b3-b4aa-2ff1b6739d34 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.631030] env[62952]: DEBUG nova.compute.provider_tree [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.633866] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 993.633866] env[62952]: value = "task-1263608" [ 993.633866] env[62952]: _type = "Task" [ 993.633866] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.643385] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263608, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.722400] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263605, 'name': ReconfigVM_Task, 'duration_secs': 0.885068} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.722735] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Reconfigured VM instance instance-00000053 to attach disk [datastore2] 6f0b04f5-811f-4c53-808b-6d9d22100a86/6f0b04f5-811f-4c53-808b-6d9d22100a86.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 993.723402] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f63cbe3f-3703-4ff9-9e9a-513b21ea822b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.730701] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 993.730701] env[62952]: value = "task-1263609" [ 993.730701] env[62952]: _type = "Task" [ 993.730701] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.740499] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263609, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.783143] env[62952]: DEBUG nova.compute.manager [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 993.852958] env[62952]: DEBUG oslo_vmware.api [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Task: {'id': task-1263607, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.697944} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.853367] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 993.853590] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 993.853827] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 993.854079] env[62952]: INFO nova.compute.manager [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Took 1.64 seconds to destroy the instance on the hypervisor. [ 993.854480] env[62952]: DEBUG oslo.service.loopingcall [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 993.854713] env[62952]: DEBUG nova.compute.manager [-] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 993.854817] env[62952]: DEBUG nova.network.neutron [-] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 994.032836] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed560501-629a-4e47-8f2e-8ab21c529f73 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.057201] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 994.057580] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d3eb2af8-e302-4753-bf52-624d88f9f56c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.067384] env[62952]: DEBUG oslo_vmware.api [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 994.067384] env[62952]: value = "task-1263610" [ 994.067384] env[62952]: _type = "Task" [ 994.067384] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.079815] env[62952]: DEBUG oslo_vmware.api [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263610, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.136394] env[62952]: DEBUG nova.scheduler.client.report [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 994.153111] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263608, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.241597] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263609, 'name': Rename_Task, 'duration_secs': 0.161543} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.241922] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 994.242229] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fd2fcba1-9c62-4f86-981b-ce0cc7f74678 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.250280] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 994.250280] env[62952]: value = "task-1263611" [ 994.250280] env[62952]: _type = "Task" [ 994.250280] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.261451] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263611, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.310843] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.324448] env[62952]: DEBUG nova.compute.manager [req-8b40c5c0-5679-4088-9b01-fbc21940ef26 req-1a47da6f-fa58-4c74-84cb-53b13cde94dd service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Received event network-vif-deleted-b5b040c9-55a9-4929-b720-a1988f7492c9 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 994.324694] env[62952]: INFO nova.compute.manager [req-8b40c5c0-5679-4088-9b01-fbc21940ef26 req-1a47da6f-fa58-4c74-84cb-53b13cde94dd service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Neutron deleted interface b5b040c9-55a9-4929-b720-a1988f7492c9; detaching it from the instance and deleting it from the info cache [ 994.324880] env[62952]: DEBUG nova.network.neutron [req-8b40c5c0-5679-4088-9b01-fbc21940ef26 req-1a47da6f-fa58-4c74-84cb-53b13cde94dd service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.582956] env[62952]: DEBUG nova.compute.manager [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Stashing vm_state: active {{(pid=62952) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 994.594485] env[62952]: DEBUG oslo_vmware.api [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263610, 'name': PowerOffVM_Task, 'duration_secs': 0.237751} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.595009] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 994.600378] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Reconfiguring VM instance instance-0000004d to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 994.600940] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8bd0d77e-b242-452b-b5b6-29cbe7ef0e6e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.621151] env[62952]: DEBUG oslo_vmware.api [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 994.621151] env[62952]: value = "task-1263612" [ 994.621151] env[62952]: _type = "Task" [ 994.621151] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.634123] env[62952]: DEBUG oslo_vmware.api [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263612, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.645609] env[62952]: DEBUG oslo_concurrency.lockutils [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.776s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.648768] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.979s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.650092] env[62952]: INFO nova.compute.claims [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 994.659176] env[62952]: DEBUG oslo_vmware.api [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263608, 'name': PowerOnVM_Task, 'duration_secs': 0.530818} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.659995] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 994.659995] env[62952]: INFO nova.compute.manager [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Took 9.57 seconds to spawn the instance on the hypervisor. [ 994.659995] env[62952]: DEBUG nova.compute.manager [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 994.660654] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48efd999-3932-4034-89cd-8939ee091582 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.673068] env[62952]: INFO nova.scheduler.client.report [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Deleted allocations for instance fba2e955-8b5d-461c-997b-ef8f3ec0d906 [ 994.684576] env[62952]: DEBUG nova.network.neutron [-] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.761413] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263611, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.828183] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-71a59169-1e3e-4e66-9fa7-20b021d1e742 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.839350] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d15268ca-0df8-4277-b718-487f6b7f6057 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.870755] env[62952]: DEBUG nova.compute.manager [req-8b40c5c0-5679-4088-9b01-fbc21940ef26 req-1a47da6f-fa58-4c74-84cb-53b13cde94dd service nova] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Detach interface failed, port_id=b5b040c9-55a9-4929-b720-a1988f7492c9, reason: Instance 0533ee81-b69d-4f4e-9832-78c5ab4512b1 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 995.118151] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.132170] env[62952]: DEBUG oslo_vmware.api [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263612, 'name': ReconfigVM_Task, 'duration_secs': 0.436042} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.132413] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Reconfigured VM instance instance-0000004d to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 995.132803] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 995.133255] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3b76a905-6c46-4554-a05d-a55e746f831e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.141975] env[62952]: DEBUG oslo_vmware.api [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 995.141975] env[62952]: value = "task-1263613" [ 995.141975] env[62952]: _type = "Task" [ 995.141975] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.152335] env[62952]: DEBUG oslo_vmware.api [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263613, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.184954] env[62952]: DEBUG oslo_concurrency.lockutils [None req-256e7f68-d664-4483-8929-dbf815bbf724 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "fba2e955-8b5d-461c-997b-ef8f3ec0d906" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.269s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.186301] env[62952]: INFO nova.compute.manager [-] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Took 1.33 seconds to deallocate network for instance. [ 995.205158] env[62952]: INFO nova.compute.manager [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Took 33.49 seconds to build instance. [ 995.262391] env[62952]: DEBUG oslo_vmware.api [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263611, 'name': PowerOnVM_Task, 'duration_secs': 0.733022} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.262737] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 995.262953] env[62952]: INFO nova.compute.manager [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Took 7.77 seconds to spawn the instance on the hypervisor. [ 995.263161] env[62952]: DEBUG nova.compute.manager [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 995.263996] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4642193d-d11a-419b-9f35-200d51219996 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.653145] env[62952]: DEBUG oslo_vmware.api [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263613, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.695499] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.707983] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa334849-fb35-4907-a6df-497536d529dd tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "7ba1587d-edaf-4dce-a224-6b2ac22cfecd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.998s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.794965] env[62952]: INFO nova.compute.manager [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Took 20.14 seconds to build instance. [ 995.878186] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad9f2237-06dc-47b1-bdf8-3fba8d8d0d50 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.887676] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b7f894-9caa-4a72-8544-3d45485f0d28 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.921411] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d81c90-3fd8-4030-a124-8dbeb76b1fae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.930357] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adeffe78-0d70-42d4-9d42-7ee19fed02a0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.945886] env[62952]: DEBUG nova.compute.provider_tree [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.045132] env[62952]: INFO nova.compute.manager [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Rescuing [ 996.045441] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "refresh_cache-7ba1587d-edaf-4dce-a224-6b2ac22cfecd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.045600] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquired lock "refresh_cache-7ba1587d-edaf-4dce-a224-6b2ac22cfecd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.045770] env[62952]: DEBUG nova.network.neutron [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 996.153941] env[62952]: DEBUG oslo_vmware.api [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263613, 'name': PowerOnVM_Task, 'duration_secs': 0.778952} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.154468] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 996.154854] env[62952]: DEBUG nova.compute.manager [None req-ef0b3902-d1d8-4735-a04d-cd67bdb938b9 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 996.155835] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b4ef11-2d0a-41b2-bfe1-5feed93196b3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.297449] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b92ee1cf-aadd-45d5-8e51-74be5ae6a9a4 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "6f0b04f5-811f-4c53-808b-6d9d22100a86" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.647s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.451212] env[62952]: DEBUG nova.scheduler.client.report [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.865701] env[62952]: DEBUG nova.network.neutron [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Updating instance_info_cache with network_info: [{"id": "a6b1b43b-91ae-45fe-8e98-bd3a194bd05b", "address": "fa:16:3e:6b:a5:77", "network": {"id": "d84b900e-d793-4187-aaeb-bff9b040e607", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1370737473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a6fddef5a2114acb9749b1bae171cbca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6b1b43b-91", "ovs_interfaceid": "a6b1b43b-91ae-45fe-8e98-bd3a194bd05b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.957224] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.309s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.957803] env[62952]: DEBUG nova.compute.manager [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 996.960782] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.546s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.961035] env[62952]: DEBUG nova.objects.instance [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lazy-loading 'resources' on Instance uuid c7d08234-fcef-4cdb-939f-05ea9e8b08ea {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.369666] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Releasing lock "refresh_cache-7ba1587d-edaf-4dce-a224-6b2ac22cfecd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.469381] env[62952]: DEBUG nova.compute.utils [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 997.474052] env[62952]: DEBUG nova.compute.manager [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 997.474277] env[62952]: DEBUG nova.network.neutron [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 997.521990] env[62952]: DEBUG nova.policy [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '469eae9db3d14e9f90988e3093514e69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff928791a3cf496cb8db2c2a6c47a6a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 997.586333] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "afaecd3c-da24-4111-8262-d92a57a42a55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.586977] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "afaecd3c-da24-4111-8262-d92a57a42a55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.714450] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c7eb0b4-8ff5-400f-8191-9e0c084bf389 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.723660] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2876184f-a24b-42e8-8c88-1c5e024d5269 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.758782] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c1e6d76-6048-4db3-8a6d-d8b4ee74d0fd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.768096] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e1b20d9-0c6d-43a9-88b1-1623349e1d42 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.783808] env[62952]: DEBUG nova.compute.provider_tree [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.868304] env[62952]: DEBUG nova.network.neutron [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Successfully created port: b0b497d8-449d-43f5-9f1a-ff2172ae47db {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 997.906642] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 997.907068] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e46c377c-7c32-4da1-848c-8c3701752e0b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.915867] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 997.915867] env[62952]: value = "task-1263614" [ 997.915867] env[62952]: _type = "Task" [ 997.915867] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.926284] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263614, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.976840] env[62952]: DEBUG nova.compute.manager [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 998.091715] env[62952]: DEBUG nova.compute.manager [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 998.269067] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquiring lock "120c6804-5aab-4406-9c0c-46eaa835b59f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.269335] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "120c6804-5aab-4406-9c0c-46eaa835b59f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.289223] env[62952]: DEBUG nova.scheduler.client.report [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.424326] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Acquiring lock "511b995d-c531-41f5-9d1b-89ec8c1bfc18" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.424599] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Lock "511b995d-c531-41f5-9d1b-89ec8c1bfc18" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.429774] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263614, 'name': PowerOffVM_Task, 'duration_secs': 0.237858} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.430049] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 998.430851] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789f9ba9-3b6d-4d0e-9fb1-5ae5aecdaef9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.454020] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8399bf-bc9d-4581-97c9-f338441d4895 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.492405] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 998.492501] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-746f18ca-0de3-4af1-b8d4-984803511c2d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.501322] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 998.501322] env[62952]: value = "task-1263615" [ 998.501322] env[62952]: _type = "Task" [ 998.501322] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.510708] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263615, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.617185] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.772208] env[62952]: DEBUG nova.compute.manager [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 998.794128] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.833s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.797036] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.486s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.798586] env[62952]: INFO nova.compute.claims [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 998.853057] env[62952]: INFO nova.scheduler.client.report [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Deleted allocations for instance c7d08234-fcef-4cdb-939f-05ea9e8b08ea [ 998.926586] env[62952]: DEBUG nova.compute.manager [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 998.988584] env[62952]: DEBUG nova.compute.manager [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 999.015422] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 999.015649] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 999.015900] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.016078] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.016273] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 999.021028] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6574c322-0d9a-426c-86e0-b7801a161971 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.024032] env[62952]: DEBUG nova.virt.hardware [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 999.024282] env[62952]: DEBUG nova.virt.hardware [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 999.024483] env[62952]: DEBUG nova.virt.hardware [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 999.024730] env[62952]: DEBUG nova.virt.hardware [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 999.024928] env[62952]: DEBUG nova.virt.hardware [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 999.025446] env[62952]: DEBUG nova.virt.hardware [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 999.025730] env[62952]: DEBUG nova.virt.hardware [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 999.025905] env[62952]: DEBUG nova.virt.hardware [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 999.026119] env[62952]: DEBUG nova.virt.hardware [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 999.026354] env[62952]: DEBUG nova.virt.hardware [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 999.026595] env[62952]: DEBUG nova.virt.hardware [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 999.027507] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4e590b-7424-44b8-8df2-c7e59e57c67b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.031812] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 999.031992] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 999.033348] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d30e45c9-7ceb-4bb6-a78c-c29871c88b93 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.039714] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61174b68-7d35-4d03-a3c9-3caeff14ca90 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.045359] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 999.045359] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5263edee-47f0-99d4-d8c3-1880d6989802" [ 999.045359] env[62952]: _type = "Task" [ 999.045359] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.065019] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5263edee-47f0-99d4-d8c3-1880d6989802, 'name': SearchDatastore_Task, 'duration_secs': 0.017358} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.066117] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c77ef8ef-613b-4e4b-a44a-139cdfea485c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.074057] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 999.074057] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524a92e0-86a5-0397-6d67-37cf34ce2ecb" [ 999.074057] env[62952]: _type = "Task" [ 999.074057] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.080799] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524a92e0-86a5-0397-6d67-37cf34ce2ecb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.301300] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.363185] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e179225-3b35-40f8-aa35-27006cf03153 tempest-ImagesTestJSON-755222348 tempest-ImagesTestJSON-755222348-project-member] Lock "c7d08234-fcef-4cdb-939f-05ea9e8b08ea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.966s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.438692] env[62952]: DEBUG nova.compute.manager [req-6339488b-4b72-49f1-8947-a7fa04442c71 req-4dc786e9-03a8-437b-b08d-ab53bcea87dd service nova] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Received event network-vif-plugged-b0b497d8-449d-43f5-9f1a-ff2172ae47db {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 999.438921] env[62952]: DEBUG oslo_concurrency.lockutils [req-6339488b-4b72-49f1-8947-a7fa04442c71 req-4dc786e9-03a8-437b-b08d-ab53bcea87dd service nova] Acquiring lock "063d1d92-0526-48e0-b33c-f51a79e7c821-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.439282] env[62952]: DEBUG oslo_concurrency.lockutils [req-6339488b-4b72-49f1-8947-a7fa04442c71 req-4dc786e9-03a8-437b-b08d-ab53bcea87dd service nova] Lock "063d1d92-0526-48e0-b33c-f51a79e7c821-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.439334] env[62952]: DEBUG oslo_concurrency.lockutils [req-6339488b-4b72-49f1-8947-a7fa04442c71 req-4dc786e9-03a8-437b-b08d-ab53bcea87dd service nova] Lock "063d1d92-0526-48e0-b33c-f51a79e7c821-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.439583] env[62952]: DEBUG nova.compute.manager [req-6339488b-4b72-49f1-8947-a7fa04442c71 req-4dc786e9-03a8-437b-b08d-ab53bcea87dd service nova] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] No waiting events found dispatching network-vif-plugged-b0b497d8-449d-43f5-9f1a-ff2172ae47db {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 999.439650] env[62952]: WARNING nova.compute.manager [req-6339488b-4b72-49f1-8947-a7fa04442c71 req-4dc786e9-03a8-437b-b08d-ab53bcea87dd service nova] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Received unexpected event network-vif-plugged-b0b497d8-449d-43f5-9f1a-ff2172ae47db for instance with vm_state building and task_state spawning. [ 999.454946] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.527347] env[62952]: DEBUG nova.network.neutron [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Successfully updated port: b0b497d8-449d-43f5-9f1a-ff2172ae47db {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 999.583366] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524a92e0-86a5-0397-6d67-37cf34ce2ecb, 'name': SearchDatastore_Task, 'duration_secs': 0.030375} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.583652] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.583919] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 7ba1587d-edaf-4dce-a224-6b2ac22cfecd/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk. {{(pid=62952) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 999.584218] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4cc9fd12-782e-4f71-8f6b-62854f3d01ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.592015] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 999.592015] env[62952]: value = "task-1263616" [ 999.592015] env[62952]: _type = "Task" [ 999.592015] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.600039] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263616, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.031373] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "refresh_cache-063d1d92-0526-48e0-b33c-f51a79e7c821" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.031373] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "refresh_cache-063d1d92-0526-48e0-b33c-f51a79e7c821" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.031373] env[62952]: DEBUG nova.network.neutron [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1000.039037] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e695ee19-a4a8-40f7-8bbf-8fdf73ab750b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.048351] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27d3562-6b96-44c2-aba2-b5bebd41f061 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.085202] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d63186e-9b79-41e7-88eb-366bd6139dba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.099359] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb8e26a-ea88-42e2-8bf9-924587bc8746 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.110133] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263616, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486732} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.117875] env[62952]: INFO nova.virt.vmwareapi.ds_util [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 7ba1587d-edaf-4dce-a224-6b2ac22cfecd/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk. [ 1000.118585] env[62952]: DEBUG nova.compute.provider_tree [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.120520] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ade24e-1adb-4e29-9cc9-4129d7327f2d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.147103] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 7ba1587d-edaf-4dce-a224-6b2ac22cfecd/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1000.147998] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b73ac27-8a03-4e1f-ba2e-c1fea8d790b3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.169200] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1000.169200] env[62952]: value = "task-1263617" [ 1000.169200] env[62952]: _type = "Task" [ 1000.169200] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.177956] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263617, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.581141] env[62952]: DEBUG nova.network.neutron [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1000.625468] env[62952]: DEBUG nova.scheduler.client.report [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1000.634350] env[62952]: DEBUG oslo_concurrency.lockutils [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "1941365e-d562-4d99-89eb-226a07e52071" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.634483] env[62952]: DEBUG oslo_concurrency.lockutils [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "1941365e-d562-4d99-89eb-226a07e52071" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.681817] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263617, 'name': ReconfigVM_Task, 'duration_secs': 0.292062} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.683109] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 7ba1587d-edaf-4dce-a224-6b2ac22cfecd/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1000.685058] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b0fa96-eda8-4eff-8f41-ed48efd14044 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.719276] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea89431c-c0fc-48d7-9af0-1b953ca7d8f8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.739240] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1000.739240] env[62952]: value = "task-1263618" [ 1000.739240] env[62952]: _type = "Task" [ 1000.739240] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.749753] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263618, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.036822] env[62952]: DEBUG nova.network.neutron [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Updating instance_info_cache with network_info: [{"id": "b0b497d8-449d-43f5-9f1a-ff2172ae47db", "address": "fa:16:3e:f3:4f:7b", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0b497d8-44", "ovs_interfaceid": "b0b497d8-449d-43f5-9f1a-ff2172ae47db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.135047] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.135790] env[62952]: DEBUG nova.compute.manager [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1001.139152] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.021s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.140693] env[62952]: DEBUG nova.compute.manager [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1001.253644] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263618, 'name': ReconfigVM_Task, 'duration_secs': 0.161609} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.253644] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1001.253644] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3cde9f66-d829-4929-b7d6-d603c39bba6c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.260566] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1001.260566] env[62952]: value = "task-1263619" [ 1001.260566] env[62952]: _type = "Task" [ 1001.260566] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.268327] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263619, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.461153] env[62952]: DEBUG nova.compute.manager [req-5e617bf3-a842-43be-91d7-9507a014b27a req-37cf5b19-c6b3-44d0-b227-e9bbb137456b service nova] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Received event network-changed-b0b497d8-449d-43f5-9f1a-ff2172ae47db {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1001.461359] env[62952]: DEBUG nova.compute.manager [req-5e617bf3-a842-43be-91d7-9507a014b27a req-37cf5b19-c6b3-44d0-b227-e9bbb137456b service nova] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Refreshing instance network info cache due to event network-changed-b0b497d8-449d-43f5-9f1a-ff2172ae47db. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1001.461566] env[62952]: DEBUG oslo_concurrency.lockutils [req-5e617bf3-a842-43be-91d7-9507a014b27a req-37cf5b19-c6b3-44d0-b227-e9bbb137456b service nova] Acquiring lock "refresh_cache-063d1d92-0526-48e0-b33c-f51a79e7c821" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.541543] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "refresh_cache-063d1d92-0526-48e0-b33c-f51a79e7c821" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.541784] env[62952]: DEBUG nova.compute.manager [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Instance network_info: |[{"id": "b0b497d8-449d-43f5-9f1a-ff2172ae47db", "address": "fa:16:3e:f3:4f:7b", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0b497d8-44", "ovs_interfaceid": "b0b497d8-449d-43f5-9f1a-ff2172ae47db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1001.542640] env[62952]: DEBUG oslo_concurrency.lockutils [req-5e617bf3-a842-43be-91d7-9507a014b27a req-37cf5b19-c6b3-44d0-b227-e9bbb137456b service nova] Acquired lock "refresh_cache-063d1d92-0526-48e0-b33c-f51a79e7c821" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.542871] env[62952]: DEBUG nova.network.neutron [req-5e617bf3-a842-43be-91d7-9507a014b27a req-37cf5b19-c6b3-44d0-b227-e9bbb137456b service nova] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Refreshing network info cache for port b0b497d8-449d-43f5-9f1a-ff2172ae47db {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1001.544259] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:4f:7b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b0b497d8-449d-43f5-9f1a-ff2172ae47db', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1001.560576] env[62952]: DEBUG oslo.service.loopingcall [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.560576] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1001.561355] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1bb06f42-f01f-4b73-a6f1-fe987ba7e244 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.587493] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1001.587493] env[62952]: value = "task-1263620" [ 1001.587493] env[62952]: _type = "Task" [ 1001.587493] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.597306] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263620, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.645151] env[62952]: DEBUG nova.compute.utils [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1001.656816] env[62952]: INFO nova.compute.claims [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1001.665433] env[62952]: DEBUG nova.compute.manager [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1001.665642] env[62952]: DEBUG nova.network.neutron [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1001.689733] env[62952]: DEBUG oslo_concurrency.lockutils [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.753303] env[62952]: DEBUG nova.policy [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e872d861b56e435d8c1250dbb82a5dfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8960909bb80f4395bb87b6de3c60ab2c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1001.771887] env[62952]: DEBUG oslo_vmware.api [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263619, 'name': PowerOnVM_Task, 'duration_secs': 0.446471} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.776042] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1001.779331] env[62952]: DEBUG nova.compute.manager [None req-eca19158-0ec3-4bdd-a543-4bfebc2f97e2 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1001.780178] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac1968d5-e04c-476d-a640-82802c069cde {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.909827] env[62952]: DEBUG nova.network.neutron [req-5e617bf3-a842-43be-91d7-9507a014b27a req-37cf5b19-c6b3-44d0-b227-e9bbb137456b service nova] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Updated VIF entry in instance network info cache for port b0b497d8-449d-43f5-9f1a-ff2172ae47db. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1001.911546] env[62952]: DEBUG nova.network.neutron [req-5e617bf3-a842-43be-91d7-9507a014b27a req-37cf5b19-c6b3-44d0-b227-e9bbb137456b service nova] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Updating instance_info_cache with network_info: [{"id": "b0b497d8-449d-43f5-9f1a-ff2172ae47db", "address": "fa:16:3e:f3:4f:7b", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0b497d8-44", "ovs_interfaceid": "b0b497d8-449d-43f5-9f1a-ff2172ae47db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.103709] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263620, 'name': CreateVM_Task, 'duration_secs': 0.466492} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.104558] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1002.105588] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.105927] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.106425] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1002.107102] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e21c5663-29f7-4bf6-8837-8ca41bcd4131 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.109864] env[62952]: DEBUG nova.network.neutron [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Successfully created port: a45070ed-ba02-43f8-b335-7cb961a6ad6d {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1002.119018] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1002.119018] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f26b77-c897-eef2-5d15-01396787584f" [ 1002.119018] env[62952]: _type = "Task" [ 1002.119018] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.128203] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f26b77-c897-eef2-5d15-01396787584f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.172879] env[62952]: DEBUG nova.compute.manager [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1002.176436] env[62952]: INFO nova.compute.resource_tracker [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Updating resource usage from migration 51bfa6dd-bc3f-4b5c-89c3-d0da48e18b27 [ 1002.415631] env[62952]: DEBUG oslo_concurrency.lockutils [req-5e617bf3-a842-43be-91d7-9507a014b27a req-37cf5b19-c6b3-44d0-b227-e9bbb137456b service nova] Releasing lock "refresh_cache-063d1d92-0526-48e0-b33c-f51a79e7c821" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.422858] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-042b5728-6bce-4225-9a01-ac2b10e33bd9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.432688] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55ee987e-2563-4c62-b32e-28e6b28673cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.466990] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d65a4a5d-1263-456f-ae3e-bdd30cb6f385 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.476935] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1c228ca-7f09-4435-91cf-06c02a47e92a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.491203] env[62952]: DEBUG nova.compute.provider_tree [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.628022] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f26b77-c897-eef2-5d15-01396787584f, 'name': SearchDatastore_Task, 'duration_secs': 0.031917} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.628194] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.628442] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1002.628755] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.629067] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.629216] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1002.629602] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1f06973c-4b53-4705-8288-8cf50569d6da {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.642706] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1002.642706] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1002.643371] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a03544a1-ccf3-48b7-afce-5c94607807b9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.649430] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1002.649430] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523ecdec-a49d-3e12-0722-9243b57fa79d" [ 1002.649430] env[62952]: _type = "Task" [ 1002.649430] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.658639] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523ecdec-a49d-3e12-0722-9243b57fa79d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.994873] env[62952]: DEBUG nova.scheduler.client.report [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.160693] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523ecdec-a49d-3e12-0722-9243b57fa79d, 'name': SearchDatastore_Task, 'duration_secs': 0.00911} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.161303] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25b8b928-f64c-433c-83af-e762751f9977 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.169268] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1003.169268] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5251bf5e-c706-fff8-0016-d220e6bda152" [ 1003.169268] env[62952]: _type = "Task" [ 1003.169268] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.176202] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5251bf5e-c706-fff8-0016-d220e6bda152, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.188292] env[62952]: DEBUG nova.compute.manager [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1003.213149] env[62952]: DEBUG nova.virt.hardware [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1003.213149] env[62952]: DEBUG nova.virt.hardware [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1003.213382] env[62952]: DEBUG nova.virt.hardware [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1003.213528] env[62952]: DEBUG nova.virt.hardware [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1003.213695] env[62952]: DEBUG nova.virt.hardware [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1003.213852] env[62952]: DEBUG nova.virt.hardware [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1003.214072] env[62952]: DEBUG nova.virt.hardware [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1003.214254] env[62952]: DEBUG nova.virt.hardware [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1003.214425] env[62952]: DEBUG nova.virt.hardware [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1003.214586] env[62952]: DEBUG nova.virt.hardware [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1003.215501] env[62952]: DEBUG nova.virt.hardware [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1003.215727] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d7b620-af37-4e42-ada7-f9f90051b901 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.225027] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc50835-17b8-42be-9e29-6a44e50a5804 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.501208] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.362s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.501555] env[62952]: INFO nova.compute.manager [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Migrating [ 1003.501910] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.502925] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.504636] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.810s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.504946] env[62952]: DEBUG nova.objects.instance [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lazy-loading 'resources' on Instance uuid 0533ee81-b69d-4f4e-9832-78c5ab4512b1 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.588820] env[62952]: DEBUG nova.compute.manager [req-14c44b6b-14b9-4401-93aa-90e1a92c5938 req-3088f2df-8199-43ce-ba4c-eedd9ae875d4 service nova] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Received event network-vif-plugged-a45070ed-ba02-43f8-b335-7cb961a6ad6d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1003.589112] env[62952]: DEBUG oslo_concurrency.lockutils [req-14c44b6b-14b9-4401-93aa-90e1a92c5938 req-3088f2df-8199-43ce-ba4c-eedd9ae875d4 service nova] Acquiring lock "e44914cd-d8ae-4f43-9207-32cacf6d6553-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.589339] env[62952]: DEBUG oslo_concurrency.lockutils [req-14c44b6b-14b9-4401-93aa-90e1a92c5938 req-3088f2df-8199-43ce-ba4c-eedd9ae875d4 service nova] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.589516] env[62952]: DEBUG oslo_concurrency.lockutils [req-14c44b6b-14b9-4401-93aa-90e1a92c5938 req-3088f2df-8199-43ce-ba4c-eedd9ae875d4 service nova] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.589699] env[62952]: DEBUG nova.compute.manager [req-14c44b6b-14b9-4401-93aa-90e1a92c5938 req-3088f2df-8199-43ce-ba4c-eedd9ae875d4 service nova] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] No waiting events found dispatching network-vif-plugged-a45070ed-ba02-43f8-b335-7cb961a6ad6d {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1003.589905] env[62952]: WARNING nova.compute.manager [req-14c44b6b-14b9-4401-93aa-90e1a92c5938 req-3088f2df-8199-43ce-ba4c-eedd9ae875d4 service nova] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Received unexpected event network-vif-plugged-a45070ed-ba02-43f8-b335-7cb961a6ad6d for instance with vm_state building and task_state spawning. [ 1003.657284] env[62952]: DEBUG nova.network.neutron [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Successfully updated port: a45070ed-ba02-43f8-b335-7cb961a6ad6d {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1003.679332] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5251bf5e-c706-fff8-0016-d220e6bda152, 'name': SearchDatastore_Task, 'duration_secs': 0.011161} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.679605] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.679866] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 063d1d92-0526-48e0-b33c-f51a79e7c821/063d1d92-0526-48e0-b33c-f51a79e7c821.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1003.680144] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-58bf4724-a671-4714-8497-fb7c6f6a415e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.688292] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1003.688292] env[62952]: value = "task-1263621" [ 1003.688292] env[62952]: _type = "Task" [ 1003.688292] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.697228] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263621, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.891689] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "d98a88aa-be5f-4254-aaba-c6fa03c9269a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.891939] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "d98a88aa-be5f-4254-aaba-c6fa03c9269a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.011668] env[62952]: INFO nova.compute.rpcapi [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 1004.011849] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.162788] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "refresh_cache-e44914cd-d8ae-4f43-9207-32cacf6d6553" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.163253] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquired lock "refresh_cache-e44914cd-d8ae-4f43-9207-32cacf6d6553" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.163513] env[62952]: DEBUG nova.network.neutron [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1004.201987] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263621, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.265427] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af68ce5-45d3-468a-8635-39d6902ea157 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.273589] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45678cd8-63a1-41d6-bb21-2235f925c466 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.306718] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be096847-475b-433d-8855-b6c9abb3bd14 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.315575] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3c6f8f-6fbe-426a-8355-58e16403393c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.330667] env[62952]: DEBUG nova.compute.provider_tree [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.394798] env[62952]: DEBUG nova.compute.manager [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1004.535490] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "refresh_cache-75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.535682] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "refresh_cache-75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.535869] env[62952]: DEBUG nova.network.neutron [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1004.702027] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263621, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.702126] env[62952]: DEBUG nova.network.neutron [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1004.834541] env[62952]: DEBUG nova.scheduler.client.report [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1004.838777] env[62952]: DEBUG nova.network.neutron [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Updating instance_info_cache with network_info: [{"id": "a45070ed-ba02-43f8-b335-7cb961a6ad6d", "address": "fa:16:3e:c3:72:0d", "network": {"id": "3c430a3e-4ea2-4c7b-8e05-733f4a306f04", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-441856160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8960909bb80f4395bb87b6de3c60ab2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa45070ed-ba", "ovs_interfaceid": "a45070ed-ba02-43f8-b335-7cb961a6ad6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.916616] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.202052] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263621, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.497834} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.202418] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 063d1d92-0526-48e0-b33c-f51a79e7c821/063d1d92-0526-48e0-b33c-f51a79e7c821.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1005.202560] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1005.203079] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-63708d43-063b-49e2-a40a-c9fd6f30012e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.210339] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1005.210339] env[62952]: value = "task-1263622" [ 1005.210339] env[62952]: _type = "Task" [ 1005.210339] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.219100] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263622, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.272243] env[62952]: DEBUG nova.network.neutron [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Updating instance_info_cache with network_info: [{"id": "ea5399c4-f18f-4c0b-82b1-062064db974b", "address": "fa:16:3e:bb:92:30", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea5399c4-f1", "ovs_interfaceid": "ea5399c4-f18f-4c0b-82b1-062064db974b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.342350] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.837s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.344635] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Releasing lock "refresh_cache-e44914cd-d8ae-4f43-9207-32cacf6d6553" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.344722] env[62952]: DEBUG nova.compute.manager [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Instance network_info: |[{"id": "a45070ed-ba02-43f8-b335-7cb961a6ad6d", "address": "fa:16:3e:c3:72:0d", "network": {"id": "3c430a3e-4ea2-4c7b-8e05-733f4a306f04", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-441856160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8960909bb80f4395bb87b6de3c60ab2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa45070ed-ba", "ovs_interfaceid": "a45070ed-ba02-43f8-b335-7cb961a6ad6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1005.345019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.728s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.346558] env[62952]: INFO nova.compute.claims [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1005.350632] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:72:0d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '03ac2c9c-6ad2-4a85-bfab-c7e336df859a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a45070ed-ba02-43f8-b335-7cb961a6ad6d', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1005.358515] env[62952]: DEBUG oslo.service.loopingcall [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1005.362015] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1005.362015] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-749dc29c-a9f4-4c79-ae33-3a4b15ad1e52 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.381346] env[62952]: INFO nova.scheduler.client.report [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Deleted allocations for instance 0533ee81-b69d-4f4e-9832-78c5ab4512b1 [ 1005.390081] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1005.390081] env[62952]: value = "task-1263623" [ 1005.390081] env[62952]: _type = "Task" [ 1005.390081] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.400333] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263623, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.631504] env[62952]: DEBUG nova.compute.manager [req-98952512-5e25-4efa-a26b-65c9a50aff9e req-aaaa723f-216f-4268-8090-5f8c51e92c7f service nova] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Received event network-changed-a45070ed-ba02-43f8-b335-7cb961a6ad6d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1005.631850] env[62952]: DEBUG nova.compute.manager [req-98952512-5e25-4efa-a26b-65c9a50aff9e req-aaaa723f-216f-4268-8090-5f8c51e92c7f service nova] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Refreshing instance network info cache due to event network-changed-a45070ed-ba02-43f8-b335-7cb961a6ad6d. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1005.632355] env[62952]: DEBUG oslo_concurrency.lockutils [req-98952512-5e25-4efa-a26b-65c9a50aff9e req-aaaa723f-216f-4268-8090-5f8c51e92c7f service nova] Acquiring lock "refresh_cache-e44914cd-d8ae-4f43-9207-32cacf6d6553" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.633538] env[62952]: DEBUG oslo_concurrency.lockutils [req-98952512-5e25-4efa-a26b-65c9a50aff9e req-aaaa723f-216f-4268-8090-5f8c51e92c7f service nova] Acquired lock "refresh_cache-e44914cd-d8ae-4f43-9207-32cacf6d6553" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.633695] env[62952]: DEBUG nova.network.neutron [req-98952512-5e25-4efa-a26b-65c9a50aff9e req-aaaa723f-216f-4268-8090-5f8c51e92c7f service nova] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Refreshing network info cache for port a45070ed-ba02-43f8-b335-7cb961a6ad6d {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1005.727288] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263622, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.24382} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.727488] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1005.728440] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eebcfeb0-4ac6-4503-9aa1-68a76a1de4cd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.758609] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 063d1d92-0526-48e0-b33c-f51a79e7c821/063d1d92-0526-48e0-b33c-f51a79e7c821.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1005.758917] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5611f0ef-b0d8-4732-a12c-fed3ed8c162a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.776215] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "refresh_cache-75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.786976] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1005.786976] env[62952]: value = "task-1263624" [ 1005.786976] env[62952]: _type = "Task" [ 1005.786976] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.800700] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263624, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.890881] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2a5be393-9c95-42e5-bff2-ff875eea5845 tempest-AttachInterfacesTestJSON-365036115 tempest-AttachInterfacesTestJSON-365036115-project-member] Lock "0533ee81-b69d-4f4e-9832-78c5ab4512b1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.688s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.902743] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263623, 'name': CreateVM_Task, 'duration_secs': 0.373965} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.902893] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1005.903581] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.903754] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.904342] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1005.904866] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-399b8de3-287b-4b43-85f3-c1f2a736dfe9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.911483] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1005.911483] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526c1ebd-a42a-1676-be53-70279ddfaf53" [ 1005.911483] env[62952]: _type = "Task" [ 1005.911483] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.921815] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526c1ebd-a42a-1676-be53-70279ddfaf53, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.300428] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263624, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.432417] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526c1ebd-a42a-1676-be53-70279ddfaf53, 'name': SearchDatastore_Task, 'duration_secs': 0.010095} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.432716] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.434174] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1006.434174] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.434174] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.434174] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1006.436374] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-27285086-caf8-47a9-b0e3-2ca6345cc19b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.450621] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1006.450832] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1006.451640] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89acf1d5-98d1-49b2-bc29-6cc4cf90f4b4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.465181] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1006.465181] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52801f83-66cc-3a81-9d04-c8d4fb6c091b" [ 1006.465181] env[62952]: _type = "Task" [ 1006.465181] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.473818] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52801f83-66cc-3a81-9d04-c8d4fb6c091b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.516611] env[62952]: DEBUG nova.network.neutron [req-98952512-5e25-4efa-a26b-65c9a50aff9e req-aaaa723f-216f-4268-8090-5f8c51e92c7f service nova] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Updated VIF entry in instance network info cache for port a45070ed-ba02-43f8-b335-7cb961a6ad6d. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1006.516611] env[62952]: DEBUG nova.network.neutron [req-98952512-5e25-4efa-a26b-65c9a50aff9e req-aaaa723f-216f-4268-8090-5f8c51e92c7f service nova] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Updating instance_info_cache with network_info: [{"id": "a45070ed-ba02-43f8-b335-7cb961a6ad6d", "address": "fa:16:3e:c3:72:0d", "network": {"id": "3c430a3e-4ea2-4c7b-8e05-733f4a306f04", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-441856160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8960909bb80f4395bb87b6de3c60ab2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa45070ed-ba", "ovs_interfaceid": "a45070ed-ba02-43f8-b335-7cb961a6ad6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.664129] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72caf48d-5575-42d8-bfdf-ea5132fd28b7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.676804] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2255079a-44e1-4acc-9236-a919778727c0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.721903] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f1e802-761e-463a-be06-6c67a227996c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.735962] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90561b2-f6f1-4174-9a93-54be206133ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.759108] env[62952]: DEBUG nova.compute.provider_tree [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.803461] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263624, 'name': ReconfigVM_Task, 'duration_secs': 0.972773} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.803780] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 063d1d92-0526-48e0-b33c-f51a79e7c821/063d1d92-0526-48e0-b33c-f51a79e7c821.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.804965] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-571c4016-93f9-4803-ae7b-51db1eb5888f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.814607] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1006.814607] env[62952]: value = "task-1263625" [ 1006.814607] env[62952]: _type = "Task" [ 1006.814607] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.825274] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263625, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.979841] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52801f83-66cc-3a81-9d04-c8d4fb6c091b, 'name': SearchDatastore_Task, 'duration_secs': 0.022854} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.979841] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d178da1e-59f7-4f8d-acf9-e230ed64efee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.985022] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1006.985022] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52421296-9b82-85a4-2833-611b0849f158" [ 1006.985022] env[62952]: _type = "Task" [ 1006.985022] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.993083] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52421296-9b82-85a4-2833-611b0849f158, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.019290] env[62952]: DEBUG oslo_concurrency.lockutils [req-98952512-5e25-4efa-a26b-65c9a50aff9e req-aaaa723f-216f-4268-8090-5f8c51e92c7f service nova] Releasing lock "refresh_cache-e44914cd-d8ae-4f43-9207-32cacf6d6553" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.260494] env[62952]: DEBUG nova.scheduler.client.report [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1007.302672] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c02d4b-9f5e-4ddd-b188-faf5bbcf2a2c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.327389] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Updating instance '75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8' progress to 0 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1007.342402] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263625, 'name': Rename_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.497623] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52421296-9b82-85a4-2833-611b0849f158, 'name': SearchDatastore_Task, 'duration_secs': 0.011419} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.497798] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.498073] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] e44914cd-d8ae-4f43-9207-32cacf6d6553/e44914cd-d8ae-4f43-9207-32cacf6d6553.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1007.498348] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c6bd24d-15c3-4b43-ab1b-3dce7322101e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.506189] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1007.506189] env[62952]: value = "task-1263626" [ 1007.506189] env[62952]: _type = "Task" [ 1007.506189] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.515562] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263626, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.669495] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Acquiring lock "4d83faf9-43db-4fc4-978b-55ca9a6ed4c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.669759] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Lock "4d83faf9-43db-4fc4-978b-55ca9a6ed4c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.769300] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.769300] env[62952]: DEBUG nova.compute.manager [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1007.770203] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.470s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.772342] env[62952]: INFO nova.compute.claims [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1007.842421] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1007.842997] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263625, 'name': Rename_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.843571] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-75397d3a-e86a-4dc8-87bb-a4ca0a017f97 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.852322] env[62952]: DEBUG oslo_vmware.api [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1007.852322] env[62952]: value = "task-1263627" [ 1007.852322] env[62952]: _type = "Task" [ 1007.852322] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.865684] env[62952]: DEBUG oslo_vmware.api [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263627, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.022914] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263626, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.172220] env[62952]: DEBUG nova.compute.manager [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1008.272465] env[62952]: DEBUG nova.compute.utils [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1008.274801] env[62952]: DEBUG nova.compute.manager [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1008.275082] env[62952]: DEBUG nova.network.neutron [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1008.323157] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "05223053-90c3-433c-99b9-e252ef9a1935" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.323582] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "05223053-90c3-433c-99b9-e252ef9a1935" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.327146] env[62952]: DEBUG nova.policy [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7eadc05f193645e6b6e86e9cdbbbaa71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '076fb3c53420447caf047f8422435152', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1008.341197] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263625, 'name': Rename_Task, 'duration_secs': 1.180104} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.341469] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1008.341743] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c7fb10dc-b471-40eb-92f7-bcc3b1542b5f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.350552] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1008.350552] env[62952]: value = "task-1263628" [ 1008.350552] env[62952]: _type = "Task" [ 1008.350552] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.365774] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263628, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.366108] env[62952]: DEBUG oslo_vmware.api [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263627, 'name': PowerOffVM_Task, 'duration_secs': 0.471955} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.366319] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1008.366502] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Updating instance '75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8' progress to 17 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1008.518635] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263626, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.90301} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.519072] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] e44914cd-d8ae-4f43-9207-32cacf6d6553/e44914cd-d8ae-4f43-9207-32cacf6d6553.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1008.519347] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1008.519661] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c86f133c-0316-4f88-b126-1a5ab52c0b1c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.528763] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1008.528763] env[62952]: value = "task-1263629" [ 1008.528763] env[62952]: _type = "Task" [ 1008.528763] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.538938] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263629, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.617533] env[62952]: DEBUG nova.network.neutron [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Successfully created port: 5242c318-dbec-47c3-b997-83b63e220ed4 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1008.706988] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.779182] env[62952]: DEBUG nova.compute.manager [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1008.826158] env[62952]: DEBUG nova.compute.manager [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1008.865986] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263628, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.876254] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1008.876976] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1008.877170] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1008.877448] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1008.877623] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1008.878263] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1008.878263] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1008.878432] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1008.878629] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1008.878839] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1008.879106] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1008.886217] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3f9a633-8092-4fdb-92c6-6f3fbadf7cd9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.905209] env[62952]: DEBUG oslo_vmware.api [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1008.905209] env[62952]: value = "task-1263630" [ 1008.905209] env[62952]: _type = "Task" [ 1008.905209] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.915023] env[62952]: DEBUG oslo_vmware.api [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263630, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.045598] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263629, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067328} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.045925] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1009.047476] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe23c1d-f5de-4a63-b3ad-7bbda8309a51 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.074025] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] e44914cd-d8ae-4f43-9207-32cacf6d6553/e44914cd-d8ae-4f43-9207-32cacf6d6553.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1009.076943] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c3c395c-ede8-4cd2-8eb8-205c954f6967 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.099122] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1009.099122] env[62952]: value = "task-1263631" [ 1009.099122] env[62952]: _type = "Task" [ 1009.099122] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.110957] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263631, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.137779] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae2483ce-9705-4459-afd8-6f55eefb7025 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.146619] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b3e0483-f259-4aa4-94f7-d8912f5ccaa5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.183544] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36df0da2-f0c6-4eaa-80ce-b6b9bde784ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.192799] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a963da7-9a0e-4d82-a0d8-5cc98d48e3e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.210276] env[62952]: DEBUG nova.compute.provider_tree [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.351831] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.363377] env[62952]: DEBUG oslo_vmware.api [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263628, 'name': PowerOnVM_Task, 'duration_secs': 0.526923} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.363656] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1009.363912] env[62952]: INFO nova.compute.manager [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Took 10.38 seconds to spawn the instance on the hypervisor. [ 1009.364267] env[62952]: DEBUG nova.compute.manager [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1009.365210] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf77293-4b58-42d5-8642-7b4592067e6d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.416921] env[62952]: DEBUG oslo_vmware.api [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263630, 'name': ReconfigVM_Task, 'duration_secs': 0.320548} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.417392] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Updating instance '75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8' progress to 33 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1009.609530] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263631, 'name': ReconfigVM_Task, 'duration_secs': 0.334359} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.609784] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Reconfigured VM instance instance-00000055 to attach disk [datastore2] e44914cd-d8ae-4f43-9207-32cacf6d6553/e44914cd-d8ae-4f43-9207-32cacf6d6553.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1009.610487] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f1da990a-5317-47b3-9361-fa0b6e5b1114 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.618206] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1009.618206] env[62952]: value = "task-1263632" [ 1009.618206] env[62952]: _type = "Task" [ 1009.618206] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.626754] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263632, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.711878] env[62952]: DEBUG nova.scheduler.client.report [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1009.795126] env[62952]: DEBUG nova.compute.manager [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1009.836155] env[62952]: DEBUG nova.virt.hardware [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1009.836155] env[62952]: DEBUG nova.virt.hardware [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1009.836748] env[62952]: DEBUG nova.virt.hardware [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1009.837322] env[62952]: DEBUG nova.virt.hardware [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1009.837667] env[62952]: DEBUG nova.virt.hardware [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1009.838015] env[62952]: DEBUG nova.virt.hardware [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1009.838551] env[62952]: DEBUG nova.virt.hardware [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1009.838921] env[62952]: DEBUG nova.virt.hardware [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1009.839321] env[62952]: DEBUG nova.virt.hardware [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1009.839685] env[62952]: DEBUG nova.virt.hardware [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1009.841018] env[62952]: DEBUG nova.virt.hardware [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1009.841703] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a7510c-1394-4477-86b4-e2f7f83ca92b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.853919] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec837c5-4de9-45c7-a1df-815a1b651dc8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.884190] env[62952]: INFO nova.compute.manager [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Took 19.23 seconds to build instance. [ 1009.926171] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1009.926461] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1009.926625] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1009.926815] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1009.926963] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1009.927148] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1009.927372] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1009.927536] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1009.927707] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1009.927873] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1009.928122] env[62952]: DEBUG nova.virt.hardware [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1009.937158] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Reconfiguring VM instance instance-00000051 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1009.937771] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec91ea07-e783-4a3d-a06e-5c7f77da7c18 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.962961] env[62952]: DEBUG oslo_vmware.api [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1009.962961] env[62952]: value = "task-1263633" [ 1009.962961] env[62952]: _type = "Task" [ 1009.962961] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.974982] env[62952]: DEBUG oslo_vmware.api [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263633, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.128438] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263632, 'name': Rename_Task, 'duration_secs': 0.215712} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.128714] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1010.129064] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c960f540-e4d7-4cb1-9ff9-7eeac098f64e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.136186] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1010.136186] env[62952]: value = "task-1263634" [ 1010.136186] env[62952]: _type = "Task" [ 1010.136186] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.145398] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263634, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.217167] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.217691] env[62952]: DEBUG nova.compute.manager [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1010.224021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.766s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.224021] env[62952]: INFO nova.compute.claims [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1010.387362] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0998f0a4-188f-46ab-b966-2e46e036c589 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "063d1d92-0526-48e0-b33c-f51a79e7c821" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.743s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.407869] env[62952]: DEBUG nova.network.neutron [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Successfully updated port: 5242c318-dbec-47c3-b997-83b63e220ed4 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1010.476086] env[62952]: DEBUG oslo_vmware.api [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263633, 'name': ReconfigVM_Task, 'duration_secs': 0.374357} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.479682] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Reconfigured VM instance instance-00000051 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1010.479682] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf3e0c13-8df3-41db-8410-bf6dae681eaf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.507647] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8/75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1010.508014] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71b65a27-11c8-4e5b-a178-5662e7f976e4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.531121] env[62952]: DEBUG oslo_vmware.api [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1010.531121] env[62952]: value = "task-1263635" [ 1010.531121] env[62952]: _type = "Task" [ 1010.531121] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.541725] env[62952]: DEBUG oslo_vmware.api [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263635, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.648946] env[62952]: DEBUG oslo_vmware.api [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263634, 'name': PowerOnVM_Task, 'duration_secs': 0.480472} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.650602] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1010.650602] env[62952]: INFO nova.compute.manager [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Took 7.46 seconds to spawn the instance on the hypervisor. [ 1010.650602] env[62952]: DEBUG nova.compute.manager [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1010.650602] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf15eed-2076-425a-833d-96f32c321a78 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.666348] env[62952]: DEBUG nova.compute.manager [req-23e7cbb8-c6df-4af8-9557-0db975256d5e req-168fb673-afd5-4a6d-bc40-6c7e02d019b5 service nova] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Received event network-vif-plugged-5242c318-dbec-47c3-b997-83b63e220ed4 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1010.666600] env[62952]: DEBUG oslo_concurrency.lockutils [req-23e7cbb8-c6df-4af8-9557-0db975256d5e req-168fb673-afd5-4a6d-bc40-6c7e02d019b5 service nova] Acquiring lock "afaecd3c-da24-4111-8262-d92a57a42a55-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.666821] env[62952]: DEBUG oslo_concurrency.lockutils [req-23e7cbb8-c6df-4af8-9557-0db975256d5e req-168fb673-afd5-4a6d-bc40-6c7e02d019b5 service nova] Lock "afaecd3c-da24-4111-8262-d92a57a42a55-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.666979] env[62952]: DEBUG oslo_concurrency.lockutils [req-23e7cbb8-c6df-4af8-9557-0db975256d5e req-168fb673-afd5-4a6d-bc40-6c7e02d019b5 service nova] Lock "afaecd3c-da24-4111-8262-d92a57a42a55-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.667177] env[62952]: DEBUG nova.compute.manager [req-23e7cbb8-c6df-4af8-9557-0db975256d5e req-168fb673-afd5-4a6d-bc40-6c7e02d019b5 service nova] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] No waiting events found dispatching network-vif-plugged-5242c318-dbec-47c3-b997-83b63e220ed4 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1010.667354] env[62952]: WARNING nova.compute.manager [req-23e7cbb8-c6df-4af8-9557-0db975256d5e req-168fb673-afd5-4a6d-bc40-6c7e02d019b5 service nova] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Received unexpected event network-vif-plugged-5242c318-dbec-47c3-b997-83b63e220ed4 for instance with vm_state building and task_state spawning. [ 1010.725828] env[62952]: DEBUG nova.compute.utils [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1010.729427] env[62952]: DEBUG nova.compute.manager [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1010.729427] env[62952]: DEBUG nova.network.neutron [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1010.796986] env[62952]: DEBUG nova.policy [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88572e088e184d70b2cb51633539ec49', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '26f601d669ce42dca2f875ca3a68525c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1010.916291] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "refresh_cache-afaecd3c-da24-4111-8262-d92a57a42a55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.916291] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "refresh_cache-afaecd3c-da24-4111-8262-d92a57a42a55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.916291] env[62952]: DEBUG nova.network.neutron [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1011.047664] env[62952]: DEBUG oslo_vmware.api [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263635, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.179741] env[62952]: INFO nova.compute.manager [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Took 16.89 seconds to build instance. [ 1011.233505] env[62952]: DEBUG nova.compute.manager [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1011.245183] env[62952]: DEBUG nova.network.neutron [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Successfully created port: a8f57522-7ac2-454d-9bac-0be6f7897f13 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1011.477048] env[62952]: DEBUG nova.network.neutron [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1011.543510] env[62952]: DEBUG oslo_vmware.api [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263635, 'name': ReconfigVM_Task, 'duration_secs': 0.616116} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.543863] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8/75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1011.544094] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Updating instance '75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8' progress to 50 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1011.550414] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "063d1d92-0526-48e0-b33c-f51a79e7c821" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.550414] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "063d1d92-0526-48e0-b33c-f51a79e7c821" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.550414] env[62952]: INFO nova.compute.manager [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Shelving [ 1011.553229] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7263c799-8670-4c06-a6a0-5900e6b048ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.563622] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d5a306e-03bf-4cfb-b879-dc60fd2882af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.603078] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dfcb24f-9d73-40a0-a1b3-89326654a9c4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.612641] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d037bc20-2311-4e09-8dbd-dc173f363e72 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.629503] env[62952]: DEBUG nova.compute.provider_tree [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.680694] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbd68c73-d92d-48b6-8208-0da748ef8d61 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.401s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.803425] env[62952]: DEBUG nova.network.neutron [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Updating instance_info_cache with network_info: [{"id": "5242c318-dbec-47c3-b997-83b63e220ed4", "address": "fa:16:3e:9c:6c:17", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5242c318-db", "ovs_interfaceid": "5242c318-dbec-47c3-b997-83b63e220ed4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.805117] env[62952]: DEBUG nova.network.neutron [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Successfully created port: 0a105c0f-99dd-4c52-b3b4-72de6af13865 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1012.060420] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1012.061512] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a0de44a-2a53-42a5-9e82-15f04ee2b9a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.068950] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52fad961-65d8-4fb5-95b0-f4b709dd3c15 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.098546] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1012.098546] env[62952]: value = "task-1263636" [ 1012.098546] env[62952]: _type = "Task" [ 1012.098546] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.100269] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b24a1c48-458d-4b1e-8d2e-6eaa6d328e65 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.116612] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263636, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.131489] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Updating instance '75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8' progress to 67 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1012.140308] env[62952]: DEBUG nova.scheduler.client.report [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.248547] env[62952]: DEBUG nova.compute.manager [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1012.288099] env[62952]: DEBUG nova.virt.hardware [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1012.288388] env[62952]: DEBUG nova.virt.hardware [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1012.288553] env[62952]: DEBUG nova.virt.hardware [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1012.290347] env[62952]: DEBUG nova.virt.hardware [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1012.290526] env[62952]: DEBUG nova.virt.hardware [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1012.290714] env[62952]: DEBUG nova.virt.hardware [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1012.290911] env[62952]: DEBUG nova.virt.hardware [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1012.291094] env[62952]: DEBUG nova.virt.hardware [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1012.291284] env[62952]: DEBUG nova.virt.hardware [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1012.291460] env[62952]: DEBUG nova.virt.hardware [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1012.291642] env[62952]: DEBUG nova.virt.hardware [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1012.292553] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-001f96d8-40d3-4239-a2c7-99c782e9b301 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.302796] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e44b01d-2cda-47ad-be83-84cbdb7e581e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.307620] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "refresh_cache-afaecd3c-da24-4111-8262-d92a57a42a55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.310032] env[62952]: DEBUG nova.compute.manager [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Instance network_info: |[{"id": "5242c318-dbec-47c3-b997-83b63e220ed4", "address": "fa:16:3e:9c:6c:17", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5242c318-db", "ovs_interfaceid": "5242c318-dbec-47c3-b997-83b63e220ed4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1012.310032] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:6c:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '365ac5b1-6d83-4dfe-887f-60574d7f6124', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5242c318-dbec-47c3-b997-83b63e220ed4', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1012.316473] env[62952]: DEBUG oslo.service.loopingcall [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1012.326980] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1012.331017] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-90ec0de7-5e8d-4cb7-9560-20823fd5fc16 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.351366] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1012.351366] env[62952]: value = "task-1263637" [ 1012.351366] env[62952]: _type = "Task" [ 1012.351366] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.359860] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263637, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.436984] env[62952]: DEBUG nova.compute.manager [req-2cd176e7-3c29-48cb-8983-7cc5acffe9ab req-40c52432-a31b-4491-94b6-50676fe00140 service nova] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Received event network-changed-a45070ed-ba02-43f8-b335-7cb961a6ad6d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1012.437272] env[62952]: DEBUG nova.compute.manager [req-2cd176e7-3c29-48cb-8983-7cc5acffe9ab req-40c52432-a31b-4491-94b6-50676fe00140 service nova] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Refreshing instance network info cache due to event network-changed-a45070ed-ba02-43f8-b335-7cb961a6ad6d. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1012.437609] env[62952]: DEBUG oslo_concurrency.lockutils [req-2cd176e7-3c29-48cb-8983-7cc5acffe9ab req-40c52432-a31b-4491-94b6-50676fe00140 service nova] Acquiring lock "refresh_cache-e44914cd-d8ae-4f43-9207-32cacf6d6553" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.437849] env[62952]: DEBUG oslo_concurrency.lockutils [req-2cd176e7-3c29-48cb-8983-7cc5acffe9ab req-40c52432-a31b-4491-94b6-50676fe00140 service nova] Acquired lock "refresh_cache-e44914cd-d8ae-4f43-9207-32cacf6d6553" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.437956] env[62952]: DEBUG nova.network.neutron [req-2cd176e7-3c29-48cb-8983-7cc5acffe9ab req-40c52432-a31b-4491-94b6-50676fe00140 service nova] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Refreshing network info cache for port a45070ed-ba02-43f8-b335-7cb961a6ad6d {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1012.611613] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263636, 'name': PowerOffVM_Task, 'duration_secs': 0.351305} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.611613] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1012.613533] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d6d87c-3beb-4ba1-a3b6-22f3584cf431 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.639047] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b21fb877-1333-43d0-81c8-c3c572df2622 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.645403] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.646034] env[62952]: DEBUG nova.compute.manager [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1012.655363] env[62952]: DEBUG oslo_concurrency.lockutils [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.966s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.656982] env[62952]: INFO nova.compute.claims [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1012.696699] env[62952]: DEBUG nova.compute.manager [req-9f2b22d7-6449-4645-a5c4-2df811c614b1 req-f635a8b3-1d0a-44bf-976a-ea6b917aad91 service nova] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Received event network-changed-5242c318-dbec-47c3-b997-83b63e220ed4 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1012.696899] env[62952]: DEBUG nova.compute.manager [req-9f2b22d7-6449-4645-a5c4-2df811c614b1 req-f635a8b3-1d0a-44bf-976a-ea6b917aad91 service nova] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Refreshing instance network info cache due to event network-changed-5242c318-dbec-47c3-b997-83b63e220ed4. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1012.697209] env[62952]: DEBUG oslo_concurrency.lockutils [req-9f2b22d7-6449-4645-a5c4-2df811c614b1 req-f635a8b3-1d0a-44bf-976a-ea6b917aad91 service nova] Acquiring lock "refresh_cache-afaecd3c-da24-4111-8262-d92a57a42a55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.697365] env[62952]: DEBUG oslo_concurrency.lockutils [req-9f2b22d7-6449-4645-a5c4-2df811c614b1 req-f635a8b3-1d0a-44bf-976a-ea6b917aad91 service nova] Acquired lock "refresh_cache-afaecd3c-da24-4111-8262-d92a57a42a55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.697538] env[62952]: DEBUG nova.network.neutron [req-9f2b22d7-6449-4645-a5c4-2df811c614b1 req-f635a8b3-1d0a-44bf-976a-ea6b917aad91 service nova] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Refreshing network info cache for port 5242c318-dbec-47c3-b997-83b63e220ed4 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1012.730048] env[62952]: DEBUG nova.network.neutron [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Port ea5399c4-f18f-4c0b-82b1-062064db974b binding to destination host cpu-1 is already ACTIVE {{(pid=62952) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1012.862811] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263637, 'name': CreateVM_Task, 'duration_secs': 0.46412} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.863206] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1012.864152] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.864463] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.864885] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1012.865314] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a74b6cb-9e5e-4e9c-a0f4-88f1c1fa48f4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.871420] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1012.871420] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a1e02b-d22e-c9ce-70ba-936095eda5b0" [ 1012.871420] env[62952]: _type = "Task" [ 1012.871420] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.881577] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a1e02b-d22e-c9ce-70ba-936095eda5b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.155343] env[62952]: DEBUG nova.compute.utils [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1013.156900] env[62952]: DEBUG nova.compute.manager [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1013.157194] env[62952]: DEBUG nova.network.neutron [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1013.169225] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1013.169225] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f0f30734-9763-4d14-afc1-57025e5efd46 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.180299] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1013.180299] env[62952]: value = "task-1263638" [ 1013.180299] env[62952]: _type = "Task" [ 1013.180299] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.193498] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263638, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.201094] env[62952]: DEBUG nova.network.neutron [req-2cd176e7-3c29-48cb-8983-7cc5acffe9ab req-40c52432-a31b-4491-94b6-50676fe00140 service nova] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Updated VIF entry in instance network info cache for port a45070ed-ba02-43f8-b335-7cb961a6ad6d. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1013.201484] env[62952]: DEBUG nova.network.neutron [req-2cd176e7-3c29-48cb-8983-7cc5acffe9ab req-40c52432-a31b-4491-94b6-50676fe00140 service nova] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Updating instance_info_cache with network_info: [{"id": "a45070ed-ba02-43f8-b335-7cb961a6ad6d", "address": "fa:16:3e:c3:72:0d", "network": {"id": "3c430a3e-4ea2-4c7b-8e05-733f4a306f04", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-441856160-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8960909bb80f4395bb87b6de3c60ab2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa45070ed-ba", "ovs_interfaceid": "a45070ed-ba02-43f8-b335-7cb961a6ad6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.254425] env[62952]: DEBUG nova.policy [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ce147db93dc492c9dafde6cd7e71d84', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a5f9a6ccf8d40a1b5ff56befd026a46', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1013.390591] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a1e02b-d22e-c9ce-70ba-936095eda5b0, 'name': SearchDatastore_Task, 'duration_secs': 0.010531} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.393202] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.393450] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1013.393674] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.393830] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.394031] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1013.394353] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c3c9d0a0-69fc-4917-b617-91bd5de6b880 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.407175] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1013.407253] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1013.408055] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8d5a9b4-b597-402a-a288-875e705864dc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.416320] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1013.416320] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5260e466-b409-0e0f-1b8e-be81c39442dd" [ 1013.416320] env[62952]: _type = "Task" [ 1013.416320] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.427027] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5260e466-b409-0e0f-1b8e-be81c39442dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.570289] env[62952]: DEBUG nova.network.neutron [req-9f2b22d7-6449-4645-a5c4-2df811c614b1 req-f635a8b3-1d0a-44bf-976a-ea6b917aad91 service nova] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Updated VIF entry in instance network info cache for port 5242c318-dbec-47c3-b997-83b63e220ed4. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1013.570289] env[62952]: DEBUG nova.network.neutron [req-9f2b22d7-6449-4645-a5c4-2df811c614b1 req-f635a8b3-1d0a-44bf-976a-ea6b917aad91 service nova] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Updating instance_info_cache with network_info: [{"id": "5242c318-dbec-47c3-b997-83b63e220ed4", "address": "fa:16:3e:9c:6c:17", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5242c318-db", "ovs_interfaceid": "5242c318-dbec-47c3-b997-83b63e220ed4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.610577] env[62952]: DEBUG nova.network.neutron [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Successfully updated port: a8f57522-7ac2-454d-9bac-0be6f7897f13 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1013.661091] env[62952]: DEBUG nova.compute.manager [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1013.691365] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263638, 'name': CreateSnapshot_Task, 'duration_secs': 0.413874} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.691748] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1013.692671] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d71637-2de6-4eed-89f1-f9a5e4b262ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.707697] env[62952]: DEBUG nova.network.neutron [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Successfully created port: a11e8e2c-69be-436f-af5b-48f53b0495c8 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1013.709463] env[62952]: DEBUG oslo_concurrency.lockutils [req-2cd176e7-3c29-48cb-8983-7cc5acffe9ab req-40c52432-a31b-4491-94b6-50676fe00140 service nova] Releasing lock "refresh_cache-e44914cd-d8ae-4f43-9207-32cacf6d6553" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.757930] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.758200] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.758430] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.930381] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5260e466-b409-0e0f-1b8e-be81c39442dd, 'name': SearchDatastore_Task, 'duration_secs': 0.010567} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.931481] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-767daa5a-108b-4090-81ab-d3cfab261812 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.936643] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e2e721-91e4-4355-9f73-ddca599f0810 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.940582] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1013.940582] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]527f7468-44ab-1243-b331-c4be9aaee049" [ 1013.940582] env[62952]: _type = "Task" [ 1013.940582] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.946836] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052ee964-add9-4718-be77-1548dadf7e9e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.953668] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]527f7468-44ab-1243-b331-c4be9aaee049, 'name': SearchDatastore_Task, 'duration_secs': 0.012119} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.954179] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.954534] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] afaecd3c-da24-4111-8262-d92a57a42a55/afaecd3c-da24-4111-8262-d92a57a42a55.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1013.984708] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce8213c0-da6e-4957-b047-88c695f4120b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.987985] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8762738-8bd3-4b7a-b047-8ba4a2712702 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.010741] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1014.010741] env[62952]: value = "task-1263642" [ 1014.010741] env[62952]: _type = "Task" [ 1014.010741] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.017417] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa21101-5f19-4f68-b39a-b00c54a48285 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.031779] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263642, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.040439] env[62952]: DEBUG nova.compute.provider_tree [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.071252] env[62952]: DEBUG oslo_concurrency.lockutils [req-9f2b22d7-6449-4645-a5c4-2df811c614b1 req-f635a8b3-1d0a-44bf-976a-ea6b917aad91 service nova] Releasing lock "refresh_cache-afaecd3c-da24-4111-8262-d92a57a42a55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.221124] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1014.222528] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-805a67a5-1ead-4296-aed6-3947fd378049 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.235717] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1014.235717] env[62952]: value = "task-1263643" [ 1014.235717] env[62952]: _type = "Task" [ 1014.235717] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.247394] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263643, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.524735] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263642, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521138} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.525137] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] afaecd3c-da24-4111-8262-d92a57a42a55/afaecd3c-da24-4111-8262-d92a57a42a55.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1014.529309] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1014.529309] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3dfa8c13-591a-4687-a85d-a9661d927faf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.539722] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1014.539722] env[62952]: value = "task-1263644" [ 1014.539722] env[62952]: _type = "Task" [ 1014.539722] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.543250] env[62952]: DEBUG nova.scheduler.client.report [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.549278] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263644, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.676522] env[62952]: DEBUG nova.compute.manager [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1014.704545] env[62952]: DEBUG nova.virt.hardware [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1014.704886] env[62952]: DEBUG nova.virt.hardware [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1014.704886] env[62952]: DEBUG nova.virt.hardware [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1014.705103] env[62952]: DEBUG nova.virt.hardware [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1014.705250] env[62952]: DEBUG nova.virt.hardware [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1014.705451] env[62952]: DEBUG nova.virt.hardware [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1014.705672] env[62952]: DEBUG nova.virt.hardware [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1014.705833] env[62952]: DEBUG nova.virt.hardware [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1014.706035] env[62952]: DEBUG nova.virt.hardware [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1014.706173] env[62952]: DEBUG nova.virt.hardware [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1014.706347] env[62952]: DEBUG nova.virt.hardware [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1014.707644] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132293ad-bf3c-47dd-a112-3e621c140520 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.716325] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cba2bb5-63a4-4d31-96e8-9f17572b4ffc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.733497] env[62952]: DEBUG nova.compute.manager [req-5bfa4425-9c09-4afe-b0eb-d0dd184e022e req-3f52f586-6f65-4470-aca7-14868ec1b863 service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Received event network-vif-plugged-a8f57522-7ac2-454d-9bac-0be6f7897f13 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1014.733497] env[62952]: DEBUG oslo_concurrency.lockutils [req-5bfa4425-9c09-4afe-b0eb-d0dd184e022e req-3f52f586-6f65-4470-aca7-14868ec1b863 service nova] Acquiring lock "120c6804-5aab-4406-9c0c-46eaa835b59f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.733497] env[62952]: DEBUG oslo_concurrency.lockutils [req-5bfa4425-9c09-4afe-b0eb-d0dd184e022e req-3f52f586-6f65-4470-aca7-14868ec1b863 service nova] Lock "120c6804-5aab-4406-9c0c-46eaa835b59f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.733497] env[62952]: DEBUG oslo_concurrency.lockutils [req-5bfa4425-9c09-4afe-b0eb-d0dd184e022e req-3f52f586-6f65-4470-aca7-14868ec1b863 service nova] Lock "120c6804-5aab-4406-9c0c-46eaa835b59f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.733957] env[62952]: DEBUG nova.compute.manager [req-5bfa4425-9c09-4afe-b0eb-d0dd184e022e req-3f52f586-6f65-4470-aca7-14868ec1b863 service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] No waiting events found dispatching network-vif-plugged-a8f57522-7ac2-454d-9bac-0be6f7897f13 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1014.733957] env[62952]: WARNING nova.compute.manager [req-5bfa4425-9c09-4afe-b0eb-d0dd184e022e req-3f52f586-6f65-4470-aca7-14868ec1b863 service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Received unexpected event network-vif-plugged-a8f57522-7ac2-454d-9bac-0be6f7897f13 for instance with vm_state building and task_state spawning. [ 1014.734188] env[62952]: DEBUG nova.compute.manager [req-5bfa4425-9c09-4afe-b0eb-d0dd184e022e req-3f52f586-6f65-4470-aca7-14868ec1b863 service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Received event network-changed-a8f57522-7ac2-454d-9bac-0be6f7897f13 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1014.734331] env[62952]: DEBUG nova.compute.manager [req-5bfa4425-9c09-4afe-b0eb-d0dd184e022e req-3f52f586-6f65-4470-aca7-14868ec1b863 service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Refreshing instance network info cache due to event network-changed-a8f57522-7ac2-454d-9bac-0be6f7897f13. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1014.734525] env[62952]: DEBUG oslo_concurrency.lockutils [req-5bfa4425-9c09-4afe-b0eb-d0dd184e022e req-3f52f586-6f65-4470-aca7-14868ec1b863 service nova] Acquiring lock "refresh_cache-120c6804-5aab-4406-9c0c-46eaa835b59f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.734666] env[62952]: DEBUG oslo_concurrency.lockutils [req-5bfa4425-9c09-4afe-b0eb-d0dd184e022e req-3f52f586-6f65-4470-aca7-14868ec1b863 service nova] Acquired lock "refresh_cache-120c6804-5aab-4406-9c0c-46eaa835b59f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.734819] env[62952]: DEBUG nova.network.neutron [req-5bfa4425-9c09-4afe-b0eb-d0dd184e022e req-3f52f586-6f65-4470-aca7-14868ec1b863 service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Refreshing network info cache for port a8f57522-7ac2-454d-9bac-0be6f7897f13 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1014.747918] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263643, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.794948] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "refresh_cache-75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.795194] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "refresh_cache-75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.795444] env[62952]: DEBUG nova.network.neutron [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1015.051048] env[62952]: DEBUG oslo_concurrency.lockutils [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.051393] env[62952]: DEBUG nova.compute.manager [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1015.054124] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263644, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.2973} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.054601] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.138s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.056201] env[62952]: INFO nova.compute.claims [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1015.058673] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1015.059833] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18a2bd4f-9e31-49e3-995e-5cc2b15121fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.084782] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] afaecd3c-da24-4111-8262-d92a57a42a55/afaecd3c-da24-4111-8262-d92a57a42a55.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1015.085160] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-481fb1f4-f0fd-4f11-8d65-184bbcf9077e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.107898] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1015.107898] env[62952]: value = "task-1263645" [ 1015.107898] env[62952]: _type = "Task" [ 1015.107898] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.117885] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263645, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.225192] env[62952]: DEBUG nova.compute.manager [req-c118cb97-3b68-4477-b0f4-7301f4bf6f3a req-bca395c5-bffb-4500-903f-1b32e8245aa4 service nova] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Received event network-vif-plugged-a11e8e2c-69be-436f-af5b-48f53b0495c8 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1015.225192] env[62952]: DEBUG oslo_concurrency.lockutils [req-c118cb97-3b68-4477-b0f4-7301f4bf6f3a req-bca395c5-bffb-4500-903f-1b32e8245aa4 service nova] Acquiring lock "511b995d-c531-41f5-9d1b-89ec8c1bfc18-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.225192] env[62952]: DEBUG oslo_concurrency.lockutils [req-c118cb97-3b68-4477-b0f4-7301f4bf6f3a req-bca395c5-bffb-4500-903f-1b32e8245aa4 service nova] Lock "511b995d-c531-41f5-9d1b-89ec8c1bfc18-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.225786] env[62952]: DEBUG oslo_concurrency.lockutils [req-c118cb97-3b68-4477-b0f4-7301f4bf6f3a req-bca395c5-bffb-4500-903f-1b32e8245aa4 service nova] Lock "511b995d-c531-41f5-9d1b-89ec8c1bfc18-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.226215] env[62952]: DEBUG nova.compute.manager [req-c118cb97-3b68-4477-b0f4-7301f4bf6f3a req-bca395c5-bffb-4500-903f-1b32e8245aa4 service nova] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] No waiting events found dispatching network-vif-plugged-a11e8e2c-69be-436f-af5b-48f53b0495c8 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1015.226574] env[62952]: WARNING nova.compute.manager [req-c118cb97-3b68-4477-b0f4-7301f4bf6f3a req-bca395c5-bffb-4500-903f-1b32e8245aa4 service nova] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Received unexpected event network-vif-plugged-a11e8e2c-69be-436f-af5b-48f53b0495c8 for instance with vm_state building and task_state spawning. [ 1015.251843] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263643, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.274718] env[62952]: DEBUG nova.network.neutron [req-5bfa4425-9c09-4afe-b0eb-d0dd184e022e req-3f52f586-6f65-4470-aca7-14868ec1b863 service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1015.365886] env[62952]: DEBUG nova.network.neutron [req-5bfa4425-9c09-4afe-b0eb-d0dd184e022e req-3f52f586-6f65-4470-aca7-14868ec1b863 service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.411290] env[62952]: DEBUG nova.network.neutron [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Successfully updated port: a11e8e2c-69be-436f-af5b-48f53b0495c8 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1015.556974] env[62952]: DEBUG nova.compute.utils [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1015.559637] env[62952]: DEBUG nova.compute.manager [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1015.559920] env[62952]: DEBUG nova.network.neutron [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1015.593340] env[62952]: DEBUG nova.network.neutron [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Updating instance_info_cache with network_info: [{"id": "ea5399c4-f18f-4c0b-82b1-062064db974b", "address": "fa:16:3e:bb:92:30", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea5399c4-f1", "ovs_interfaceid": "ea5399c4-f18f-4c0b-82b1-062064db974b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.618952] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263645, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.632799] env[62952]: DEBUG nova.policy [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '23bdda22bcf74a1ea9d681512d663aeb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '091cdeb6da48492bb02f93822a45c9bf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1015.658538] env[62952]: DEBUG nova.network.neutron [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Successfully updated port: 0a105c0f-99dd-4c52-b3b4-72de6af13865 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1015.748111] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263643, 'name': CloneVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.868966] env[62952]: DEBUG oslo_concurrency.lockutils [req-5bfa4425-9c09-4afe-b0eb-d0dd184e022e req-3f52f586-6f65-4470-aca7-14868ec1b863 service nova] Releasing lock "refresh_cache-120c6804-5aab-4406-9c0c-46eaa835b59f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.915455] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Acquiring lock "refresh_cache-511b995d-c531-41f5-9d1b-89ec8c1bfc18" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.915755] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Acquired lock "refresh_cache-511b995d-c531-41f5-9d1b-89ec8c1bfc18" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.915966] env[62952]: DEBUG nova.network.neutron [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1016.072551] env[62952]: DEBUG nova.compute.manager [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1016.098221] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "refresh_cache-75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.120993] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263645, 'name': ReconfigVM_Task, 'duration_secs': 0.808476} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.121570] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Reconfigured VM instance instance-00000056 to attach disk [datastore1] afaecd3c-da24-4111-8262-d92a57a42a55/afaecd3c-da24-4111-8262-d92a57a42a55.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1016.122386] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4a46df35-b1a6-463c-b6ba-ced3c2a3b1a7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.134151] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1016.134151] env[62952]: value = "task-1263646" [ 1016.134151] env[62952]: _type = "Task" [ 1016.134151] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.146573] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263646, 'name': Rename_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.164989] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquiring lock "refresh_cache-120c6804-5aab-4406-9c0c-46eaa835b59f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.169258] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquired lock "refresh_cache-120c6804-5aab-4406-9c0c-46eaa835b59f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.169258] env[62952]: DEBUG nova.network.neutron [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1016.214294] env[62952]: DEBUG nova.network.neutron [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Successfully created port: e1c235af-9501-47c0-a83a-ba40ea7fa8c1 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1016.251167] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263643, 'name': CloneVM_Task, 'duration_secs': 1.561889} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.253861] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Created linked-clone VM from snapshot [ 1016.255596] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78180295-818a-40d0-9baa-fb83a3f92bfd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.266168] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Uploading image 13c9a867-a743-44b0-87b1-abaa7ed74f6c {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1016.297257] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1016.297257] env[62952]: value = "vm-271943" [ 1016.297257] env[62952]: _type = "VirtualMachine" [ 1016.297257] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1016.297590] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a2b27271-523f-47c2-b591-9bf506725e81 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.308358] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lease: (returnval){ [ 1016.308358] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f30c8b-df42-4770-a847-265f7f875645" [ 1016.308358] env[62952]: _type = "HttpNfcLease" [ 1016.308358] env[62952]: } obtained for exporting VM: (result){ [ 1016.308358] env[62952]: value = "vm-271943" [ 1016.308358] env[62952]: _type = "VirtualMachine" [ 1016.308358] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1016.309227] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the lease: (returnval){ [ 1016.309227] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f30c8b-df42-4770-a847-265f7f875645" [ 1016.309227] env[62952]: _type = "HttpNfcLease" [ 1016.309227] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1016.320043] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1016.320043] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f30c8b-df42-4770-a847-265f7f875645" [ 1016.320043] env[62952]: _type = "HttpNfcLease" [ 1016.320043] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1016.367074] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1334f436-5a99-473c-a82b-fdaa32480afd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.389518] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf4dc42-11fd-4806-b35a-080b4eea6e9e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.435961] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-705d607c-6285-48d6-a566-7510a9111417 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.447111] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a8f9cc-8c48-42b9-bb9c-5720e8fcf20b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.467489] env[62952]: DEBUG nova.compute.provider_tree [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1016.477260] env[62952]: DEBUG nova.network.neutron [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1016.619460] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcd1630b-6401-410e-83ee-c2c1f09cda70 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.640056] env[62952]: DEBUG nova.network.neutron [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Updating instance_info_cache with network_info: [{"id": "a11e8e2c-69be-436f-af5b-48f53b0495c8", "address": "fa:16:3e:0b:39:21", "network": {"id": "b2f625f1-cadb-4173-8115-7b809cf8a7f1", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-220503173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a5f9a6ccf8d40a1b5ff56befd026a46", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "447ff42d-b33e-4b5d-8b7f-e8117ebbbc92", "external-id": "nsx-vlan-transportzone-836", "segmentation_id": 836, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa11e8e2c-69", "ovs_interfaceid": "a11e8e2c-69be-436f-af5b-48f53b0495c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.644610] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd852853-03d2-418f-9616-d62280d69fb6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.657563] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263646, 'name': Rename_Task, 'duration_secs': 0.154001} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.657807] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Updating instance '75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8' progress to 83 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1016.661484] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1016.662202] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-554cf0bc-6d69-4b2d-a602-39ec785bb5cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.669403] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1016.669403] env[62952]: value = "task-1263649" [ 1016.669403] env[62952]: _type = "Task" [ 1016.669403] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.680732] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263649, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.706582] env[62952]: DEBUG nova.network.neutron [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1016.758664] env[62952]: DEBUG nova.compute.manager [req-202f415a-dbb9-48dc-a4c2-cf1aee7d4cf8 req-d3d79239-2f0d-4c38-a196-02664bda3c0b service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Received event network-vif-plugged-0a105c0f-99dd-4c52-b3b4-72de6af13865 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.758996] env[62952]: DEBUG oslo_concurrency.lockutils [req-202f415a-dbb9-48dc-a4c2-cf1aee7d4cf8 req-d3d79239-2f0d-4c38-a196-02664bda3c0b service nova] Acquiring lock "120c6804-5aab-4406-9c0c-46eaa835b59f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.758996] env[62952]: DEBUG oslo_concurrency.lockutils [req-202f415a-dbb9-48dc-a4c2-cf1aee7d4cf8 req-d3d79239-2f0d-4c38-a196-02664bda3c0b service nova] Lock "120c6804-5aab-4406-9c0c-46eaa835b59f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.759241] env[62952]: DEBUG oslo_concurrency.lockutils [req-202f415a-dbb9-48dc-a4c2-cf1aee7d4cf8 req-d3d79239-2f0d-4c38-a196-02664bda3c0b service nova] Lock "120c6804-5aab-4406-9c0c-46eaa835b59f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.759241] env[62952]: DEBUG nova.compute.manager [req-202f415a-dbb9-48dc-a4c2-cf1aee7d4cf8 req-d3d79239-2f0d-4c38-a196-02664bda3c0b service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] No waiting events found dispatching network-vif-plugged-0a105c0f-99dd-4c52-b3b4-72de6af13865 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1016.759781] env[62952]: WARNING nova.compute.manager [req-202f415a-dbb9-48dc-a4c2-cf1aee7d4cf8 req-d3d79239-2f0d-4c38-a196-02664bda3c0b service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Received unexpected event network-vif-plugged-0a105c0f-99dd-4c52-b3b4-72de6af13865 for instance with vm_state building and task_state spawning. [ 1016.759781] env[62952]: DEBUG nova.compute.manager [req-202f415a-dbb9-48dc-a4c2-cf1aee7d4cf8 req-d3d79239-2f0d-4c38-a196-02664bda3c0b service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Received event network-changed-0a105c0f-99dd-4c52-b3b4-72de6af13865 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.759781] env[62952]: DEBUG nova.compute.manager [req-202f415a-dbb9-48dc-a4c2-cf1aee7d4cf8 req-d3d79239-2f0d-4c38-a196-02664bda3c0b service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Refreshing instance network info cache due to event network-changed-0a105c0f-99dd-4c52-b3b4-72de6af13865. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1016.759977] env[62952]: DEBUG oslo_concurrency.lockutils [req-202f415a-dbb9-48dc-a4c2-cf1aee7d4cf8 req-d3d79239-2f0d-4c38-a196-02664bda3c0b service nova] Acquiring lock "refresh_cache-120c6804-5aab-4406-9c0c-46eaa835b59f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.818948] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1016.818948] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f30c8b-df42-4770-a847-265f7f875645" [ 1016.818948] env[62952]: _type = "HttpNfcLease" [ 1016.818948] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1016.821909] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1016.821909] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f30c8b-df42-4770-a847-265f7f875645" [ 1016.821909] env[62952]: _type = "HttpNfcLease" [ 1016.821909] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1016.822862] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-105f0412-93d9-4c40-9ee6-a420ca22fc3c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.834432] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527ba0cc-0586-d899-029c-860737b22b35/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1016.834724] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527ba0cc-0586-d899-029c-860737b22b35/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1016.963094] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a4e13343-f90b-4875-987d-1d373adabd0a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.970663] env[62952]: DEBUG nova.scheduler.client.report [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1017.065186] env[62952]: DEBUG nova.network.neutron [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Updating instance_info_cache with network_info: [{"id": "a8f57522-7ac2-454d-9bac-0be6f7897f13", "address": "fa:16:3e:4e:c4:91", "network": {"id": "2ee3625b-dfcf-47dd-94d7-518ca3072b17", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1511607833", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26f601d669ce42dca2f875ca3a68525c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8f57522-7a", "ovs_interfaceid": "a8f57522-7ac2-454d-9bac-0be6f7897f13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0a105c0f-99dd-4c52-b3b4-72de6af13865", "address": "fa:16:3e:dc:0e:0c", "network": {"id": "1451063a-71b6-4695-a2ea-12ca546a9b45", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-264997297", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.57", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "26f601d669ce42dca2f875ca3a68525c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f625f389-b7cf-49b9-998a-87f3a9e3f234", "external-id": "nsx-vlan-transportzone-280", "segmentation_id": 280, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a105c0f-99", "ovs_interfaceid": "0a105c0f-99dd-4c52-b3b4-72de6af13865", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.081052] env[62952]: DEBUG nova.compute.manager [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1017.107222] env[62952]: DEBUG nova.virt.hardware [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1017.107566] env[62952]: DEBUG nova.virt.hardware [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1017.107755] env[62952]: DEBUG nova.virt.hardware [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1017.108553] env[62952]: DEBUG nova.virt.hardware [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1017.108553] env[62952]: DEBUG nova.virt.hardware [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1017.108553] env[62952]: DEBUG nova.virt.hardware [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1017.109329] env[62952]: DEBUG nova.virt.hardware [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1017.109329] env[62952]: DEBUG nova.virt.hardware [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1017.109329] env[62952]: DEBUG nova.virt.hardware [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1017.109329] env[62952]: DEBUG nova.virt.hardware [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1017.109470] env[62952]: DEBUG nova.virt.hardware [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1017.110293] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8087021d-673b-4477-9f84-ae43eafffe6e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.121879] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95346199-a1c2-459f-86eb-aedf1476343b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.150329] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Releasing lock "refresh_cache-511b995d-c531-41f5-9d1b-89ec8c1bfc18" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.150624] env[62952]: DEBUG nova.compute.manager [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Instance network_info: |[{"id": "a11e8e2c-69be-436f-af5b-48f53b0495c8", "address": "fa:16:3e:0b:39:21", "network": {"id": "b2f625f1-cadb-4173-8115-7b809cf8a7f1", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-220503173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a5f9a6ccf8d40a1b5ff56befd026a46", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "447ff42d-b33e-4b5d-8b7f-e8117ebbbc92", "external-id": "nsx-vlan-transportzone-836", "segmentation_id": 836, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa11e8e2c-69", "ovs_interfaceid": "a11e8e2c-69be-436f-af5b-48f53b0495c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1017.151197] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:39:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '447ff42d-b33e-4b5d-8b7f-e8117ebbbc92', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a11e8e2c-69be-436f-af5b-48f53b0495c8', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1017.159248] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Creating folder: Project (1a5f9a6ccf8d40a1b5ff56befd026a46). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1017.159568] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7bdbdedd-2005-4eb5-973a-da7430697631 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.164863] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1017.165711] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-55c56aff-473a-4585-88be-540784245294 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.175252] env[62952]: DEBUG oslo_vmware.api [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1017.175252] env[62952]: value = "task-1263651" [ 1017.175252] env[62952]: _type = "Task" [ 1017.175252] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.183060] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Created folder: Project (1a5f9a6ccf8d40a1b5ff56befd026a46) in parent group-v271811. [ 1017.183236] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Creating folder: Instances. Parent ref: group-v271944. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1017.183484] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263649, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.184980] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f6034c1a-4faf-4cf0-9a73-110701b623fb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.191749] env[62952]: DEBUG oslo_vmware.api [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263651, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.201073] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Created folder: Instances in parent group-v271944. [ 1017.201844] env[62952]: DEBUG oslo.service.loopingcall [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1017.201844] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1017.201844] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-46d7bfde-ac8e-495c-b0d7-1aa8de8d9f92 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.223270] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1017.223270] env[62952]: value = "task-1263653" [ 1017.223270] env[62952]: _type = "Task" [ 1017.223270] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.234523] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263653, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.260864] env[62952]: DEBUG nova.compute.manager [req-9d1aee24-0f7c-41cb-9a2d-46964f3f39e3 req-48bb2442-5be3-4b5d-b32d-f69a99f8bc80 service nova] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Received event network-changed-a11e8e2c-69be-436f-af5b-48f53b0495c8 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1017.261123] env[62952]: DEBUG nova.compute.manager [req-9d1aee24-0f7c-41cb-9a2d-46964f3f39e3 req-48bb2442-5be3-4b5d-b32d-f69a99f8bc80 service nova] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Refreshing instance network info cache due to event network-changed-a11e8e2c-69be-436f-af5b-48f53b0495c8. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1017.261465] env[62952]: DEBUG oslo_concurrency.lockutils [req-9d1aee24-0f7c-41cb-9a2d-46964f3f39e3 req-48bb2442-5be3-4b5d-b32d-f69a99f8bc80 service nova] Acquiring lock "refresh_cache-511b995d-c531-41f5-9d1b-89ec8c1bfc18" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.261641] env[62952]: DEBUG oslo_concurrency.lockutils [req-9d1aee24-0f7c-41cb-9a2d-46964f3f39e3 req-48bb2442-5be3-4b5d-b32d-f69a99f8bc80 service nova] Acquired lock "refresh_cache-511b995d-c531-41f5-9d1b-89ec8c1bfc18" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.262488] env[62952]: DEBUG nova.network.neutron [req-9d1aee24-0f7c-41cb-9a2d-46964f3f39e3 req-48bb2442-5be3-4b5d-b32d-f69a99f8bc80 service nova] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Refreshing network info cache for port a11e8e2c-69be-436f-af5b-48f53b0495c8 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1017.476256] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.477113] env[62952]: DEBUG nova.compute.manager [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1017.480903] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.774s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.482807] env[62952]: INFO nova.compute.claims [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1017.567976] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Releasing lock "refresh_cache-120c6804-5aab-4406-9c0c-46eaa835b59f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.568401] env[62952]: DEBUG nova.compute.manager [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Instance network_info: |[{"id": "a8f57522-7ac2-454d-9bac-0be6f7897f13", "address": "fa:16:3e:4e:c4:91", "network": {"id": "2ee3625b-dfcf-47dd-94d7-518ca3072b17", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1511607833", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26f601d669ce42dca2f875ca3a68525c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8f57522-7a", "ovs_interfaceid": "a8f57522-7ac2-454d-9bac-0be6f7897f13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0a105c0f-99dd-4c52-b3b4-72de6af13865", "address": "fa:16:3e:dc:0e:0c", "network": {"id": "1451063a-71b6-4695-a2ea-12ca546a9b45", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-264997297", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.57", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "26f601d669ce42dca2f875ca3a68525c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f625f389-b7cf-49b9-998a-87f3a9e3f234", "external-id": "nsx-vlan-transportzone-280", "segmentation_id": 280, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a105c0f-99", "ovs_interfaceid": "0a105c0f-99dd-4c52-b3b4-72de6af13865", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1017.568783] env[62952]: DEBUG oslo_concurrency.lockutils [req-202f415a-dbb9-48dc-a4c2-cf1aee7d4cf8 req-d3d79239-2f0d-4c38-a196-02664bda3c0b service nova] Acquired lock "refresh_cache-120c6804-5aab-4406-9c0c-46eaa835b59f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.569489] env[62952]: DEBUG nova.network.neutron [req-202f415a-dbb9-48dc-a4c2-cf1aee7d4cf8 req-d3d79239-2f0d-4c38-a196-02664bda3c0b service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Refreshing network info cache for port 0a105c0f-99dd-4c52-b3b4-72de6af13865 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1017.571804] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4e:c4:91', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '359850cc-b061-4c9c-a61c-eb42e0f7c359', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a8f57522-7ac2-454d-9bac-0be6f7897f13', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:0e:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f625f389-b7cf-49b9-998a-87f3a9e3f234', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0a105c0f-99dd-4c52-b3b4-72de6af13865', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1017.583992] env[62952]: DEBUG oslo.service.loopingcall [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1017.584899] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1017.585242] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-797d9083-4b34-47e7-820b-4f79aa787450 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.613554] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1017.613554] env[62952]: value = "task-1263654" [ 1017.613554] env[62952]: _type = "Task" [ 1017.613554] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.626044] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263654, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.689469] env[62952]: DEBUG oslo_vmware.api [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263649, 'name': PowerOnVM_Task, 'duration_secs': 0.820279} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.689936] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1017.690393] env[62952]: INFO nova.compute.manager [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Took 7.90 seconds to spawn the instance on the hypervisor. [ 1017.691977] env[62952]: DEBUG nova.compute.manager [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1017.698625] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19a5cc2b-336e-47cf-969c-e64f158377b5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.702166] env[62952]: DEBUG oslo_vmware.api [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263651, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.737177] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263653, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.985497] env[62952]: DEBUG nova.compute.utils [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1017.987654] env[62952]: DEBUG nova.compute.manager [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1017.987876] env[62952]: DEBUG nova.network.neutron [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1018.037571] env[62952]: DEBUG nova.network.neutron [req-9d1aee24-0f7c-41cb-9a2d-46964f3f39e3 req-48bb2442-5be3-4b5d-b32d-f69a99f8bc80 service nova] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Updated VIF entry in instance network info cache for port a11e8e2c-69be-436f-af5b-48f53b0495c8. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1018.037969] env[62952]: DEBUG nova.network.neutron [req-9d1aee24-0f7c-41cb-9a2d-46964f3f39e3 req-48bb2442-5be3-4b5d-b32d-f69a99f8bc80 service nova] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Updating instance_info_cache with network_info: [{"id": "a11e8e2c-69be-436f-af5b-48f53b0495c8", "address": "fa:16:3e:0b:39:21", "network": {"id": "b2f625f1-cadb-4173-8115-7b809cf8a7f1", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-220503173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a5f9a6ccf8d40a1b5ff56befd026a46", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "447ff42d-b33e-4b5d-8b7f-e8117ebbbc92", "external-id": "nsx-vlan-transportzone-836", "segmentation_id": 836, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa11e8e2c-69", "ovs_interfaceid": "a11e8e2c-69be-436f-af5b-48f53b0495c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.105414] env[62952]: DEBUG nova.policy [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b7e0ae49fe64e56bddf86570ed8270b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a6fddef5a2114acb9749b1bae171cbca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1018.126825] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263654, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.188192] env[62952]: DEBUG oslo_vmware.api [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263651, 'name': PowerOnVM_Task, 'duration_secs': 0.623989} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.188547] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1018.188806] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d094c168-1a73-4e5e-a1ee-84985686fa28 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Updating instance '75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8' progress to 100 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1018.220831] env[62952]: INFO nova.compute.manager [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Took 19.62 seconds to build instance. [ 1018.237445] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263653, 'name': CreateVM_Task, 'duration_secs': 0.546103} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.240343] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1018.241171] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.241343] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.241973] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1018.242538] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d417d87-f786-4d57-a4e8-1e33a7f66ffd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.249087] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Waiting for the task: (returnval){ [ 1018.249087] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520db5a9-1d91-bc49-4011-c6b52d7225f5" [ 1018.249087] env[62952]: _type = "Task" [ 1018.249087] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.262857] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520db5a9-1d91-bc49-4011-c6b52d7225f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.299210] env[62952]: DEBUG nova.network.neutron [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Successfully updated port: e1c235af-9501-47c0-a83a-ba40ea7fa8c1 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1018.317308] env[62952]: DEBUG nova.network.neutron [req-202f415a-dbb9-48dc-a4c2-cf1aee7d4cf8 req-d3d79239-2f0d-4c38-a196-02664bda3c0b service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Updated VIF entry in instance network info cache for port 0a105c0f-99dd-4c52-b3b4-72de6af13865. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1018.317308] env[62952]: DEBUG nova.network.neutron [req-202f415a-dbb9-48dc-a4c2-cf1aee7d4cf8 req-d3d79239-2f0d-4c38-a196-02664bda3c0b service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Updating instance_info_cache with network_info: [{"id": "a8f57522-7ac2-454d-9bac-0be6f7897f13", "address": "fa:16:3e:4e:c4:91", "network": {"id": "2ee3625b-dfcf-47dd-94d7-518ca3072b17", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1511607833", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26f601d669ce42dca2f875ca3a68525c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8f57522-7a", "ovs_interfaceid": "a8f57522-7ac2-454d-9bac-0be6f7897f13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0a105c0f-99dd-4c52-b3b4-72de6af13865", "address": "fa:16:3e:dc:0e:0c", "network": {"id": "1451063a-71b6-4695-a2ea-12ca546a9b45", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-264997297", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.57", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "26f601d669ce42dca2f875ca3a68525c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f625f389-b7cf-49b9-998a-87f3a9e3f234", "external-id": "nsx-vlan-transportzone-280", "segmentation_id": 280, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a105c0f-99", "ovs_interfaceid": "0a105c0f-99dd-4c52-b3b4-72de6af13865", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.395942] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.396489] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.497150] env[62952]: DEBUG nova.compute.manager [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1018.542158] env[62952]: DEBUG oslo_concurrency.lockutils [req-9d1aee24-0f7c-41cb-9a2d-46964f3f39e3 req-48bb2442-5be3-4b5d-b32d-f69a99f8bc80 service nova] Releasing lock "refresh_cache-511b995d-c531-41f5-9d1b-89ec8c1bfc18" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.631562] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263654, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.652041] env[62952]: DEBUG nova.network.neutron [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Successfully created port: 3dff9ce5-4a76-49a1-ac02-2702ccde508f {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1018.725037] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6746347-2846-45fd-af04-ad8bc434e94b tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "afaecd3c-da24-4111-8262-d92a57a42a55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.138s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.765809] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520db5a9-1d91-bc49-4011-c6b52d7225f5, 'name': SearchDatastore_Task, 'duration_secs': 0.026864} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.766536] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.766796] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.767152] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.767327] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.767583] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.767849] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-96872df0-9197-4efc-ae11-ee087adad8c7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.781462] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.781747] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1018.782626] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-284a4899-6551-4c16-9ed8-a5f035f89134 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.792212] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Waiting for the task: (returnval){ [ 1018.792212] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524cb658-1804-5017-a529-e7a43cb27d64" [ 1018.792212] env[62952]: _type = "Task" [ 1018.792212] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.800779] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524cb658-1804-5017-a529-e7a43cb27d64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.802180] env[62952]: DEBUG oslo_concurrency.lockutils [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.802233] env[62952]: DEBUG oslo_concurrency.lockutils [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.802371] env[62952]: DEBUG nova.network.neutron [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1018.803913] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c99b462-9286-4a98-9a13-ad7b1bcdb9e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.812397] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e0aa3c8-6b78-43c8-a1d6-757f9281e7e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.819551] env[62952]: DEBUG oslo_concurrency.lockutils [req-202f415a-dbb9-48dc-a4c2-cf1aee7d4cf8 req-d3d79239-2f0d-4c38-a196-02664bda3c0b service nova] Releasing lock "refresh_cache-120c6804-5aab-4406-9c0c-46eaa835b59f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.851247] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad35faba-e3c7-423f-b4f1-545a6de58256 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.861617] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b9acdb-e431-447d-9ed1-1c3a3d218af9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.877773] env[62952]: DEBUG nova.compute.provider_tree [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.900439] env[62952]: INFO nova.compute.manager [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Detaching volume 11fed6c5-a8bb-4173-ade8-22ca4aa0b85c [ 1018.955407] env[62952]: INFO nova.virt.block_device [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Attempting to driver detach volume 11fed6c5-a8bb-4173-ade8-22ca4aa0b85c from mountpoint /dev/sdb [ 1018.955642] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1018.955832] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271910', 'volume_id': '11fed6c5-a8bb-4173-ade8-22ca4aa0b85c', 'name': 'volume-11fed6c5-a8bb-4173-ade8-22ca4aa0b85c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '9cc7bc2f-ff22-43cf-be79-ff8c6426115b', 'attached_at': '', 'detached_at': '', 'volume_id': '11fed6c5-a8bb-4173-ade8-22ca4aa0b85c', 'serial': '11fed6c5-a8bb-4173-ade8-22ca4aa0b85c'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1018.957531] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a7e76b2-2ee4-49ec-8410-5696af6dbb95 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.979987] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-defdd926-95dc-4b60-b8a1-f9eb97c2c63f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.989561] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd12dc36-0397-4153-8b99-3ec0847a3d90 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.018371] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e67bedb-a71c-416b-b28c-712350db4f36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.037059] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] The volume has not been displaced from its original location: [datastore1] volume-11fed6c5-a8bb-4173-ade8-22ca4aa0b85c/volume-11fed6c5-a8bb-4173-ade8-22ca4aa0b85c.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1019.042871] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Reconfiguring VM instance instance-00000033 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1019.043939] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b539711-c271-49e1-96fb-2152d890ed86 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.063290] env[62952]: DEBUG oslo_vmware.api [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1019.063290] env[62952]: value = "task-1263656" [ 1019.063290] env[62952]: _type = "Task" [ 1019.063290] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.076572] env[62952]: DEBUG oslo_vmware.api [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263656, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.125983] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263654, 'name': CreateVM_Task, 'duration_secs': 1.378751} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.126183] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1019.127032] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.127159] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.127664] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1019.127848] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c671cbda-c4ee-4200-8a09-f636c6e55f56 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.133660] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 1019.133660] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523b0984-f665-2aa1-2be6-ff8a9555184a" [ 1019.133660] env[62952]: _type = "Task" [ 1019.133660] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.144059] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523b0984-f665-2aa1-2be6-ff8a9555184a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.304815] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524cb658-1804-5017-a529-e7a43cb27d64, 'name': SearchDatastore_Task, 'duration_secs': 0.015558} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.305760] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69d235c6-86a6-460e-a3fe-35ed9b85385e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.316672] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Waiting for the task: (returnval){ [ 1019.316672] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b38ec5-506d-c4db-07d6-57d648a9b3a7" [ 1019.316672] env[62952]: _type = "Task" [ 1019.316672] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.325932] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b38ec5-506d-c4db-07d6-57d648a9b3a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.342387] env[62952]: DEBUG nova.network.neutron [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1019.381155] env[62952]: DEBUG nova.scheduler.client.report [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.491155] env[62952]: DEBUG nova.network.neutron [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Updating instance_info_cache with network_info: [{"id": "e1c235af-9501-47c0-a83a-ba40ea7fa8c1", "address": "fa:16:3e:34:07:9b", "network": {"id": "827a8171-6759-4434-a6f1-437aad881c81", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1902486970-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091cdeb6da48492bb02f93822a45c9bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11b669be-fb26-4ef8-bdb6-c77ab9d06daf", "external-id": "nsx-vlan-transportzone-633", "segmentation_id": 633, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1c235af-95", "ovs_interfaceid": "e1c235af-9501-47c0-a83a-ba40ea7fa8c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.509784] env[62952]: DEBUG nova.compute.manager [req-e9a32a63-9226-427a-8967-aa9dc094b445 req-fa456ec5-7413-4922-9a03-9757272df832 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Received event network-vif-plugged-e1c235af-9501-47c0-a83a-ba40ea7fa8c1 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1019.509973] env[62952]: DEBUG oslo_concurrency.lockutils [req-e9a32a63-9226-427a-8967-aa9dc094b445 req-fa456ec5-7413-4922-9a03-9757272df832 service nova] Acquiring lock "1941365e-d562-4d99-89eb-226a07e52071-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.510333] env[62952]: DEBUG oslo_concurrency.lockutils [req-e9a32a63-9226-427a-8967-aa9dc094b445 req-fa456ec5-7413-4922-9a03-9757272df832 service nova] Lock "1941365e-d562-4d99-89eb-226a07e52071-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.511296] env[62952]: DEBUG oslo_concurrency.lockutils [req-e9a32a63-9226-427a-8967-aa9dc094b445 req-fa456ec5-7413-4922-9a03-9757272df832 service nova] Lock "1941365e-d562-4d99-89eb-226a07e52071-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.511296] env[62952]: DEBUG nova.compute.manager [req-e9a32a63-9226-427a-8967-aa9dc094b445 req-fa456ec5-7413-4922-9a03-9757272df832 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] No waiting events found dispatching network-vif-plugged-e1c235af-9501-47c0-a83a-ba40ea7fa8c1 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1019.511296] env[62952]: WARNING nova.compute.manager [req-e9a32a63-9226-427a-8967-aa9dc094b445 req-fa456ec5-7413-4922-9a03-9757272df832 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Received unexpected event network-vif-plugged-e1c235af-9501-47c0-a83a-ba40ea7fa8c1 for instance with vm_state building and task_state spawning. [ 1019.511296] env[62952]: DEBUG nova.compute.manager [req-e9a32a63-9226-427a-8967-aa9dc094b445 req-fa456ec5-7413-4922-9a03-9757272df832 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Received event network-changed-e1c235af-9501-47c0-a83a-ba40ea7fa8c1 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1019.511986] env[62952]: DEBUG nova.compute.manager [req-e9a32a63-9226-427a-8967-aa9dc094b445 req-fa456ec5-7413-4922-9a03-9757272df832 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Refreshing instance network info cache due to event network-changed-e1c235af-9501-47c0-a83a-ba40ea7fa8c1. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1019.511986] env[62952]: DEBUG oslo_concurrency.lockutils [req-e9a32a63-9226-427a-8967-aa9dc094b445 req-fa456ec5-7413-4922-9a03-9757272df832 service nova] Acquiring lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.524172] env[62952]: DEBUG nova.compute.manager [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1019.556855] env[62952]: DEBUG nova.virt.hardware [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1019.557537] env[62952]: DEBUG nova.virt.hardware [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1019.558274] env[62952]: DEBUG nova.virt.hardware [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1019.559889] env[62952]: DEBUG nova.virt.hardware [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1019.560190] env[62952]: DEBUG nova.virt.hardware [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1019.560481] env[62952]: DEBUG nova.virt.hardware [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1019.560935] env[62952]: DEBUG nova.virt.hardware [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1019.561305] env[62952]: DEBUG nova.virt.hardware [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1019.561628] env[62952]: DEBUG nova.virt.hardware [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1019.561975] env[62952]: DEBUG nova.virt.hardware [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1019.562289] env[62952]: DEBUG nova.virt.hardware [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1019.563867] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd67477-d6cd-42d4-881e-46affa67c6f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.579479] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c4b1374-fb01-4153-9bdf-30a636a2ca3c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.585257] env[62952]: DEBUG oslo_vmware.api [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263656, 'name': ReconfigVM_Task, 'duration_secs': 0.345046} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.585701] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Reconfigured VM instance instance-00000033 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1019.592329] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac56ce54-761b-491d-9e0b-bb3d659c5cec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.620265] env[62952]: DEBUG oslo_vmware.api [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1019.620265] env[62952]: value = "task-1263657" [ 1019.620265] env[62952]: _type = "Task" [ 1019.620265] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.629969] env[62952]: DEBUG oslo_vmware.api [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263657, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.647263] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523b0984-f665-2aa1-2be6-ff8a9555184a, 'name': SearchDatastore_Task, 'duration_secs': 0.013461} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.648426] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.648426] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1019.648426] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.828558] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b38ec5-506d-c4db-07d6-57d648a9b3a7, 'name': SearchDatastore_Task, 'duration_secs': 0.017205} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.829104] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.829263] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 511b995d-c531-41f5-9d1b-89ec8c1bfc18/511b995d-c531-41f5-9d1b-89ec8c1bfc18.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1019.829747] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.829836] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1019.830109] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-42a8b261-fb9c-4822-b776-da4a2f1d6b46 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.833624] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e7b61f98-71f2-487c-8172-63bc5f9fd8ae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.841049] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Waiting for the task: (returnval){ [ 1019.841049] env[62952]: value = "task-1263658" [ 1019.841049] env[62952]: _type = "Task" [ 1019.841049] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.846071] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1019.846310] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1019.847890] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-973289d6-8c58-4693-93ad-9fc18c8d4d2a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.854107] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263658, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.859781] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 1019.859781] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]522bdc1a-e55f-0b8e-c116-a9cb2a7a584b" [ 1019.859781] env[62952]: _type = "Task" [ 1019.859781] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.868597] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]522bdc1a-e55f-0b8e-c116-a9cb2a7a584b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.888690] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.892730] env[62952]: DEBUG nova.compute.manager [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1019.897054] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.544s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.898074] env[62952]: INFO nova.compute.claims [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1019.993161] env[62952]: DEBUG oslo_concurrency.lockutils [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Releasing lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.993672] env[62952]: DEBUG nova.compute.manager [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Instance network_info: |[{"id": "e1c235af-9501-47c0-a83a-ba40ea7fa8c1", "address": "fa:16:3e:34:07:9b", "network": {"id": "827a8171-6759-4434-a6f1-437aad881c81", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1902486970-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091cdeb6da48492bb02f93822a45c9bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11b669be-fb26-4ef8-bdb6-c77ab9d06daf", "external-id": "nsx-vlan-transportzone-633", "segmentation_id": 633, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1c235af-95", "ovs_interfaceid": "e1c235af-9501-47c0-a83a-ba40ea7fa8c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1019.994136] env[62952]: DEBUG oslo_concurrency.lockutils [req-e9a32a63-9226-427a-8967-aa9dc094b445 req-fa456ec5-7413-4922-9a03-9757272df832 service nova] Acquired lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.994604] env[62952]: DEBUG nova.network.neutron [req-e9a32a63-9226-427a-8967-aa9dc094b445 req-fa456ec5-7413-4922-9a03-9757272df832 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Refreshing network info cache for port e1c235af-9501-47c0-a83a-ba40ea7fa8c1 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1019.995763] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:34:07:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '11b669be-fb26-4ef8-bdb6-c77ab9d06daf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e1c235af-9501-47c0-a83a-ba40ea7fa8c1', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1020.005759] env[62952]: DEBUG oslo.service.loopingcall [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.010436] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1020.011174] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1d3145e6-de47-44b0-8eab-e2328e9ed2e0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.037586] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1020.037586] env[62952]: value = "task-1263659" [ 1020.037586] env[62952]: _type = "Task" [ 1020.037586] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.049594] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263659, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.132327] env[62952]: DEBUG oslo_vmware.api [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263657, 'name': ReconfigVM_Task, 'duration_secs': 0.188503} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.132666] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271910', 'volume_id': '11fed6c5-a8bb-4173-ade8-22ca4aa0b85c', 'name': 'volume-11fed6c5-a8bb-4173-ade8-22ca4aa0b85c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '9cc7bc2f-ff22-43cf-be79-ff8c6426115b', 'attached_at': '', 'detached_at': '', 'volume_id': '11fed6c5-a8bb-4173-ade8-22ca4aa0b85c', 'serial': '11fed6c5-a8bb-4173-ade8-22ca4aa0b85c'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1020.189663] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.190021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.190309] env[62952]: DEBUG nova.compute.manager [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Going to confirm migration 1 {{(pid=62952) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1020.261192] env[62952]: DEBUG nova.network.neutron [req-e9a32a63-9226-427a-8967-aa9dc094b445 req-fa456ec5-7413-4922-9a03-9757272df832 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Updated VIF entry in instance network info cache for port e1c235af-9501-47c0-a83a-ba40ea7fa8c1. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1020.261192] env[62952]: DEBUG nova.network.neutron [req-e9a32a63-9226-427a-8967-aa9dc094b445 req-fa456ec5-7413-4922-9a03-9757272df832 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Updating instance_info_cache with network_info: [{"id": "e1c235af-9501-47c0-a83a-ba40ea7fa8c1", "address": "fa:16:3e:34:07:9b", "network": {"id": "827a8171-6759-4434-a6f1-437aad881c81", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1902486970-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091cdeb6da48492bb02f93822a45c9bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11b669be-fb26-4ef8-bdb6-c77ab9d06daf", "external-id": "nsx-vlan-transportzone-633", "segmentation_id": 633, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1c235af-95", "ovs_interfaceid": "e1c235af-9501-47c0-a83a-ba40ea7fa8c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.353632] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263658, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.370979] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]522bdc1a-e55f-0b8e-c116-a9cb2a7a584b, 'name': SearchDatastore_Task, 'duration_secs': 0.01366} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.371920] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abad3f8c-fb26-48dd-ba6e-d744062aff2b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.377823] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 1020.377823] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c62a78-d32f-5d35-784b-183b39dc4c34" [ 1020.377823] env[62952]: _type = "Task" [ 1020.377823] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.386893] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c62a78-d32f-5d35-784b-183b39dc4c34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.394775] env[62952]: DEBUG nova.compute.utils [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1020.396238] env[62952]: DEBUG nova.compute.manager [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1020.396418] env[62952]: DEBUG nova.network.neutron [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1020.401896] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "afaecd3c-da24-4111-8262-d92a57a42a55" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.402171] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "afaecd3c-da24-4111-8262-d92a57a42a55" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.402380] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "afaecd3c-da24-4111-8262-d92a57a42a55-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.402560] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "afaecd3c-da24-4111-8262-d92a57a42a55-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.404252] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "afaecd3c-da24-4111-8262-d92a57a42a55-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.409492] env[62952]: INFO nova.compute.manager [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Terminating instance [ 1020.412855] env[62952]: DEBUG nova.compute.manager [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1020.416016] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1020.416016] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db190664-cdaa-4639-848c-1e10f6330396 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.424904] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1020.425228] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9910fa7a-2c65-485d-9b48-917da0a1ba10 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.433112] env[62952]: DEBUG oslo_vmware.api [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1020.433112] env[62952]: value = "task-1263660" [ 1020.433112] env[62952]: _type = "Task" [ 1020.433112] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.443896] env[62952]: DEBUG oslo_vmware.api [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263660, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.456607] env[62952]: DEBUG nova.policy [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '60ac1f7b9ff24c4ca1acd72cb061e7c6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e340d25add224233aec18cf4c7c62ac3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1020.551037] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263659, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.691352] env[62952]: DEBUG nova.objects.instance [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lazy-loading 'flavor' on Instance uuid 9cc7bc2f-ff22-43cf-be79-ff8c6426115b {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1020.763773] env[62952]: DEBUG oslo_concurrency.lockutils [req-e9a32a63-9226-427a-8967-aa9dc094b445 req-fa456ec5-7413-4922-9a03-9757272df832 service nova] Releasing lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.772045] env[62952]: DEBUG nova.network.neutron [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Successfully created port: 272ccc8b-6da5-4832-9304-bcf2814b9283 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1020.795874] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "refresh_cache-75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.796189] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "refresh_cache-75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.796387] env[62952]: DEBUG nova.network.neutron [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1020.796578] env[62952]: DEBUG nova.objects.instance [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lazy-loading 'info_cache' on Instance uuid 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1020.854529] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263658, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.889209] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c62a78-d32f-5d35-784b-183b39dc4c34, 'name': SearchDatastore_Task, 'duration_secs': 0.020095} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.889599] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.889884] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 120c6804-5aab-4406-9c0c-46eaa835b59f/120c6804-5aab-4406-9c0c-46eaa835b59f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1020.890208] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-68d8d46f-13a9-47fa-94b5-4ad1300710c7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.898261] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 1020.898261] env[62952]: value = "task-1263662" [ 1020.898261] env[62952]: _type = "Task" [ 1020.898261] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.907734] env[62952]: DEBUG nova.compute.manager [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1020.912138] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263662, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.948219] env[62952]: DEBUG oslo_vmware.api [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263660, 'name': PowerOffVM_Task, 'duration_secs': 0.434132} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.948575] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1020.948777] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1020.949047] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6a6d05a6-3cc1-4064-b032-93065bca1990 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.988640] env[62952]: DEBUG nova.network.neutron [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Successfully updated port: 3dff9ce5-4a76-49a1-ac02-2702ccde508f {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1021.049117] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263659, 'name': CreateVM_Task, 'duration_secs': 0.547809} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.051949] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1021.052898] env[62952]: DEBUG oslo_concurrency.lockutils [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.053137] env[62952]: DEBUG oslo_concurrency.lockutils [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.053508] env[62952]: DEBUG oslo_concurrency.lockutils [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1021.053776] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c88f9fc2-8451-42ee-9836-29c87e3264d0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.059934] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1021.059934] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52db27d3-75a4-135a-2ccd-9fca742baa3c" [ 1021.059934] env[62952]: _type = "Task" [ 1021.059934] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.071539] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52db27d3-75a4-135a-2ccd-9fca742baa3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.215744] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb4195b-63cc-437b-9544-3cfa817d61d0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.224955] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7043b5-030c-469b-affe-5446cf313742 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.255357] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42fb6a02-337b-432e-acdd-cf541159e4a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.263806] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babe9a4e-5ee2-49cf-9849-f6118d3ee00e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.278188] env[62952]: DEBUG nova.compute.provider_tree [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1021.352641] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263658, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.032306} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.352910] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 511b995d-c531-41f5-9d1b-89ec8c1bfc18/511b995d-c531-41f5-9d1b-89ec8c1bfc18.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1021.353142] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1021.353392] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f3abecf5-36db-4f11-99ba-2c04424bcbba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.360839] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Waiting for the task: (returnval){ [ 1021.360839] env[62952]: value = "task-1263664" [ 1021.360839] env[62952]: _type = "Task" [ 1021.360839] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.369253] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263664, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.408301] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263662, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.494153] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "refresh_cache-d98a88aa-be5f-4254-aaba-c6fa03c9269a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.494329] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquired lock "refresh_cache-d98a88aa-be5f-4254-aaba-c6fa03c9269a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.494488] env[62952]: DEBUG nova.network.neutron [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1021.573581] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52db27d3-75a4-135a-2ccd-9fca742baa3c, 'name': SearchDatastore_Task, 'duration_secs': 0.022222} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.574065] env[62952]: DEBUG oslo_concurrency.lockutils [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.574313] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1021.574570] env[62952]: DEBUG oslo_concurrency.lockutils [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.574805] env[62952]: DEBUG oslo_concurrency.lockutils [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.575275] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1021.575611] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f93de92e-8242-42c4-92f4-57f618f20954 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.586699] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1021.586882] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1021.588025] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88060609-78fc-436e-b403-8ebdb3d3fe02 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.597305] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1021.597305] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524a9118-6962-e215-0668-1b4ddfb5da4b" [ 1021.597305] env[62952]: _type = "Task" [ 1021.597305] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.605998] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524a9118-6962-e215-0668-1b4ddfb5da4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.647498] env[62952]: DEBUG nova.compute.manager [req-b0370da5-4189-442b-9cd8-6cede235b97b req-1c4d7d98-8e3b-47cc-8855-ee9236b11dc7 service nova] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Received event network-vif-plugged-3dff9ce5-4a76-49a1-ac02-2702ccde508f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1021.647842] env[62952]: DEBUG oslo_concurrency.lockutils [req-b0370da5-4189-442b-9cd8-6cede235b97b req-1c4d7d98-8e3b-47cc-8855-ee9236b11dc7 service nova] Acquiring lock "d98a88aa-be5f-4254-aaba-c6fa03c9269a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.647954] env[62952]: DEBUG oslo_concurrency.lockutils [req-b0370da5-4189-442b-9cd8-6cede235b97b req-1c4d7d98-8e3b-47cc-8855-ee9236b11dc7 service nova] Lock "d98a88aa-be5f-4254-aaba-c6fa03c9269a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.648125] env[62952]: DEBUG oslo_concurrency.lockutils [req-b0370da5-4189-442b-9cd8-6cede235b97b req-1c4d7d98-8e3b-47cc-8855-ee9236b11dc7 service nova] Lock "d98a88aa-be5f-4254-aaba-c6fa03c9269a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.648301] env[62952]: DEBUG nova.compute.manager [req-b0370da5-4189-442b-9cd8-6cede235b97b req-1c4d7d98-8e3b-47cc-8855-ee9236b11dc7 service nova] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] No waiting events found dispatching network-vif-plugged-3dff9ce5-4a76-49a1-ac02-2702ccde508f {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1021.648467] env[62952]: WARNING nova.compute.manager [req-b0370da5-4189-442b-9cd8-6cede235b97b req-1c4d7d98-8e3b-47cc-8855-ee9236b11dc7 service nova] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Received unexpected event network-vif-plugged-3dff9ce5-4a76-49a1-ac02-2702ccde508f for instance with vm_state building and task_state spawning. [ 1021.648889] env[62952]: DEBUG nova.compute.manager [req-b0370da5-4189-442b-9cd8-6cede235b97b req-1c4d7d98-8e3b-47cc-8855-ee9236b11dc7 service nova] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Received event network-changed-3dff9ce5-4a76-49a1-ac02-2702ccde508f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1021.648889] env[62952]: DEBUG nova.compute.manager [req-b0370da5-4189-442b-9cd8-6cede235b97b req-1c4d7d98-8e3b-47cc-8855-ee9236b11dc7 service nova] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Refreshing instance network info cache due to event network-changed-3dff9ce5-4a76-49a1-ac02-2702ccde508f. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1021.648993] env[62952]: DEBUG oslo_concurrency.lockutils [req-b0370da5-4189-442b-9cd8-6cede235b97b req-1c4d7d98-8e3b-47cc-8855-ee9236b11dc7 service nova] Acquiring lock "refresh_cache-d98a88aa-be5f-4254-aaba-c6fa03c9269a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.705686] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f406b0d-1e88-4155-80cd-c788887fd6e2 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.309s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.783023] env[62952]: DEBUG nova.scheduler.client.report [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1021.872857] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263664, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.220023} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.873255] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1021.874219] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab61d6c4-a99c-42a5-af7d-e3f7cb8c9ba5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.897557] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 511b995d-c531-41f5-9d1b-89ec8c1bfc18/511b995d-c531-41f5-9d1b-89ec8c1bfc18.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1021.897884] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbac15d2-150e-41eb-93e6-39190ac0aede {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.922062] env[62952]: DEBUG nova.compute.manager [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1021.926079] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Waiting for the task: (returnval){ [ 1021.926079] env[62952]: value = "task-1263665" [ 1021.926079] env[62952]: _type = "Task" [ 1021.926079] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.936057] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263662, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.922682} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.938485] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 120c6804-5aab-4406-9c0c-46eaa835b59f/120c6804-5aab-4406-9c0c-46eaa835b59f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1021.938485] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1021.938718] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1021.940887] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1021.940887] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleting the datastore file [datastore1] afaecd3c-da24-4111-8262-d92a57a42a55 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1021.940887] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c8002393-12e5-43c8-8d55-f5ff700a5c3f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.941881] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-46c8a35e-e2df-4596-93fb-b0bf419a5bea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.952907] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263665, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.960407] env[62952]: DEBUG oslo_vmware.api [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1021.960407] env[62952]: value = "task-1263667" [ 1021.960407] env[62952]: _type = "Task" [ 1021.960407] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.961969] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 1021.961969] env[62952]: value = "task-1263666" [ 1021.961969] env[62952]: _type = "Task" [ 1021.961969] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.964613] env[62952]: DEBUG nova.virt.hardware [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1021.965251] env[62952]: DEBUG nova.virt.hardware [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1021.965251] env[62952]: DEBUG nova.virt.hardware [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1021.965251] env[62952]: DEBUG nova.virt.hardware [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1021.965420] env[62952]: DEBUG nova.virt.hardware [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1021.965457] env[62952]: DEBUG nova.virt.hardware [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1021.965649] env[62952]: DEBUG nova.virt.hardware [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1021.965815] env[62952]: DEBUG nova.virt.hardware [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1021.965995] env[62952]: DEBUG nova.virt.hardware [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1021.966371] env[62952]: DEBUG nova.virt.hardware [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1021.966568] env[62952]: DEBUG nova.virt.hardware [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1021.970434] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb7953f-c539-4947-a70a-228a9e511099 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.986469] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263666, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.986801] env[62952]: DEBUG oslo_vmware.api [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263667, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.993332] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f44ff7-71e2-4165-8dbd-e1d4e431dd54 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.076495] env[62952]: DEBUG nova.network.neutron [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1022.107912] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524a9118-6962-e215-0668-1b4ddfb5da4b, 'name': SearchDatastore_Task, 'duration_secs': 0.026104} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.109836] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ffdcc5b-84eb-48bc-a5bc-b5da0c5a2efa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.117010] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1022.117010] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5225dc5e-41e2-58cd-ccaa-e28dccda2cb8" [ 1022.117010] env[62952]: _type = "Task" [ 1022.117010] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.128846] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5225dc5e-41e2-58cd-ccaa-e28dccda2cb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.236521] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.236806] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.237126] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.237447] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.237727] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.240490] env[62952]: INFO nova.compute.manager [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Terminating instance [ 1022.243065] env[62952]: DEBUG nova.compute.manager [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1022.243284] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1022.244122] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d39c64b-819e-4a19-a2d5-40c92255fe7f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.252856] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1022.253516] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fdf6943f-5cda-4535-a248-343b416d377a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.262754] env[62952]: DEBUG oslo_vmware.api [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1022.262754] env[62952]: value = "task-1263668" [ 1022.262754] env[62952]: _type = "Task" [ 1022.262754] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.274730] env[62952]: DEBUG oslo_vmware.api [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263668, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.286788] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.391s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.287339] env[62952]: DEBUG nova.compute.manager [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1022.328062] env[62952]: DEBUG nova.network.neutron [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Updating instance_info_cache with network_info: [{"id": "ea5399c4-f18f-4c0b-82b1-062064db974b", "address": "fa:16:3e:bb:92:30", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea5399c4-f1", "ovs_interfaceid": "ea5399c4-f18f-4c0b-82b1-062064db974b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.422630] env[62952]: DEBUG nova.network.neutron [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Updating instance_info_cache with network_info: [{"id": "3dff9ce5-4a76-49a1-ac02-2702ccde508f", "address": "fa:16:3e:95:0b:fe", "network": {"id": "d84b900e-d793-4187-aaeb-bff9b040e607", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1370737473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a6fddef5a2114acb9749b1bae171cbca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dff9ce5-4a", "ovs_interfaceid": "3dff9ce5-4a76-49a1-ac02-2702ccde508f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.438873] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263665, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.472422] env[62952]: DEBUG oslo_vmware.api [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263667, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.407905} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.476577] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1022.476577] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1022.476577] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1022.476713] env[62952]: INFO nova.compute.manager [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Took 2.06 seconds to destroy the instance on the hypervisor. [ 1022.476893] env[62952]: DEBUG oslo.service.loopingcall [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.477433] env[62952]: DEBUG nova.compute.manager [-] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1022.477665] env[62952]: DEBUG nova.network.neutron [-] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1022.485586] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263666, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085328} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.485924] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1022.486770] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac93949-7ac2-419c-ba5e-c86bd3586eba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.515216] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 120c6804-5aab-4406-9c0c-46eaa835b59f/120c6804-5aab-4406-9c0c-46eaa835b59f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1022.516363] env[62952]: DEBUG nova.network.neutron [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Successfully updated port: 272ccc8b-6da5-4832-9304-bcf2814b9283 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1022.520033] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9d6bb4d-ba94-44da-8b93-894beb609c97 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.549744] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 1022.549744] env[62952]: value = "task-1263669" [ 1022.549744] env[62952]: _type = "Task" [ 1022.549744] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.559771] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263669, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.626702] env[62952]: DEBUG nova.compute.manager [req-357ddd43-9265-43e4-8204-bb85f9c3cf12 req-c0057921-0c03-4806-8816-e09c0c4ad2bd service nova] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Received event network-vif-plugged-272ccc8b-6da5-4832-9304-bcf2814b9283 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.627136] env[62952]: DEBUG oslo_concurrency.lockutils [req-357ddd43-9265-43e4-8204-bb85f9c3cf12 req-c0057921-0c03-4806-8816-e09c0c4ad2bd service nova] Acquiring lock "4d83faf9-43db-4fc4-978b-55ca9a6ed4c9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.627225] env[62952]: DEBUG oslo_concurrency.lockutils [req-357ddd43-9265-43e4-8204-bb85f9c3cf12 req-c0057921-0c03-4806-8816-e09c0c4ad2bd service nova] Lock "4d83faf9-43db-4fc4-978b-55ca9a6ed4c9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.627340] env[62952]: DEBUG oslo_concurrency.lockutils [req-357ddd43-9265-43e4-8204-bb85f9c3cf12 req-c0057921-0c03-4806-8816-e09c0c4ad2bd service nova] Lock "4d83faf9-43db-4fc4-978b-55ca9a6ed4c9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.627550] env[62952]: DEBUG nova.compute.manager [req-357ddd43-9265-43e4-8204-bb85f9c3cf12 req-c0057921-0c03-4806-8816-e09c0c4ad2bd service nova] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] No waiting events found dispatching network-vif-plugged-272ccc8b-6da5-4832-9304-bcf2814b9283 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1022.627800] env[62952]: WARNING nova.compute.manager [req-357ddd43-9265-43e4-8204-bb85f9c3cf12 req-c0057921-0c03-4806-8816-e09c0c4ad2bd service nova] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Received unexpected event network-vif-plugged-272ccc8b-6da5-4832-9304-bcf2814b9283 for instance with vm_state building and task_state spawning. [ 1022.632710] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5225dc5e-41e2-58cd-ccaa-e28dccda2cb8, 'name': SearchDatastore_Task, 'duration_secs': 0.01559} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.632710] env[62952]: DEBUG oslo_concurrency.lockutils [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.632887] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 1941365e-d562-4d99-89eb-226a07e52071/1941365e-d562-4d99-89eb-226a07e52071.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1022.633242] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a094a276-5d49-489c-b34f-2d7b33e8a653 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.641746] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1022.641746] env[62952]: value = "task-1263670" [ 1022.641746] env[62952]: _type = "Task" [ 1022.641746] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.651981] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263670, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.776473] env[62952]: DEBUG oslo_vmware.api [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263668, 'name': PowerOffVM_Task, 'duration_secs': 0.33101} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.776757] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1022.776981] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1022.777649] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d24f28d1-ab6c-4500-98e4-9488c57184bc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.796767] env[62952]: DEBUG nova.compute.utils [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1022.798854] env[62952]: DEBUG nova.compute.manager [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1022.799114] env[62952]: DEBUG nova.network.neutron [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1022.832232] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "refresh_cache-75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.833056] env[62952]: DEBUG nova.objects.instance [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lazy-loading 'migration_context' on Instance uuid 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.850193] env[62952]: DEBUG nova.policy [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ffc122d64b19432aae8d9627d2928c06', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5080fbeac0744478843ab28efea8fc18', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1022.869317] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1022.869317] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1022.869317] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Deleting the datastore file [datastore1] 9cc7bc2f-ff22-43cf-be79-ff8c6426115b {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1022.869317] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bd96073e-41a2-4e2d-b7a5-6caa7a468944 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.876231] env[62952]: DEBUG oslo_vmware.api [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1022.876231] env[62952]: value = "task-1263673" [ 1022.876231] env[62952]: _type = "Task" [ 1022.876231] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.887956] env[62952]: DEBUG oslo_vmware.api [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263673, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.927950] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Releasing lock "refresh_cache-d98a88aa-be5f-4254-aaba-c6fa03c9269a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.928323] env[62952]: DEBUG nova.compute.manager [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Instance network_info: |[{"id": "3dff9ce5-4a76-49a1-ac02-2702ccde508f", "address": "fa:16:3e:95:0b:fe", "network": {"id": "d84b900e-d793-4187-aaeb-bff9b040e607", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1370737473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a6fddef5a2114acb9749b1bae171cbca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dff9ce5-4a", "ovs_interfaceid": "3dff9ce5-4a76-49a1-ac02-2702ccde508f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1022.928791] env[62952]: DEBUG oslo_concurrency.lockutils [req-b0370da5-4189-442b-9cd8-6cede235b97b req-1c4d7d98-8e3b-47cc-8855-ee9236b11dc7 service nova] Acquired lock "refresh_cache-d98a88aa-be5f-4254-aaba-c6fa03c9269a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.929061] env[62952]: DEBUG nova.network.neutron [req-b0370da5-4189-442b-9cd8-6cede235b97b req-1c4d7d98-8e3b-47cc-8855-ee9236b11dc7 service nova] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Refreshing network info cache for port 3dff9ce5-4a76-49a1-ac02-2702ccde508f {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1022.930501] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:0b:fe', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc9714ff-7109-4ea1-9435-b2b3fbdb9e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3dff9ce5-4a76-49a1-ac02-2702ccde508f', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1022.939468] env[62952]: DEBUG oslo.service.loopingcall [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.946949] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1022.947686] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aad2d843-cbe9-4141-b4dd-bfee37f05c69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.970962] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263665, 'name': ReconfigVM_Task, 'duration_secs': 0.715555} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.972680] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 511b995d-c531-41f5-9d1b-89ec8c1bfc18/511b995d-c531-41f5-9d1b-89ec8c1bfc18.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1022.973439] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1022.973439] env[62952]: value = "task-1263674" [ 1022.973439] env[62952]: _type = "Task" [ 1022.973439] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.973599] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c9399faa-6fad-4233-bb69-0786eb987215 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.985495] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263674, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.987140] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Waiting for the task: (returnval){ [ 1022.987140] env[62952]: value = "task-1263675" [ 1022.987140] env[62952]: _type = "Task" [ 1022.987140] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.998342] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263675, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.019875] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Acquiring lock "refresh_cache-4d83faf9-43db-4fc4-978b-55ca9a6ed4c9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.020531] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Acquired lock "refresh_cache-4d83faf9-43db-4fc4-978b-55ca9a6ed4c9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.020531] env[62952]: DEBUG nova.network.neutron [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1023.061423] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263669, 'name': ReconfigVM_Task, 'duration_secs': 0.361619} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.061728] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 120c6804-5aab-4406-9c0c-46eaa835b59f/120c6804-5aab-4406-9c0c-46eaa835b59f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1023.062832] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e9b3d070-7343-488b-979d-f09a4d87e4a6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.072513] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 1023.072513] env[62952]: value = "task-1263676" [ 1023.072513] env[62952]: _type = "Task" [ 1023.072513] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.086697] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263676, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.168790] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263670, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.206983] env[62952]: DEBUG nova.network.neutron [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Successfully created port: fe560683-a241-423f-9f16-c5f59668d967 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1023.308461] env[62952]: DEBUG nova.compute.manager [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1023.318963] env[62952]: DEBUG nova.network.neutron [req-b0370da5-4189-442b-9cd8-6cede235b97b req-1c4d7d98-8e3b-47cc-8855-ee9236b11dc7 service nova] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Updated VIF entry in instance network info cache for port 3dff9ce5-4a76-49a1-ac02-2702ccde508f. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1023.318963] env[62952]: DEBUG nova.network.neutron [req-b0370da5-4189-442b-9cd8-6cede235b97b req-1c4d7d98-8e3b-47cc-8855-ee9236b11dc7 service nova] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Updating instance_info_cache with network_info: [{"id": "3dff9ce5-4a76-49a1-ac02-2702ccde508f", "address": "fa:16:3e:95:0b:fe", "network": {"id": "d84b900e-d793-4187-aaeb-bff9b040e607", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1370737473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a6fddef5a2114acb9749b1bae171cbca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dff9ce5-4a", "ovs_interfaceid": "3dff9ce5-4a76-49a1-ac02-2702ccde508f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.338742] env[62952]: DEBUG nova.objects.base [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Object Instance<75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8> lazy-loaded attributes: info_cache,migration_context {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1023.340460] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc3c6b4-6911-4e1c-ad7a-0c1efaeeecc9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.376127] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f09e9186-c394-44a9-8c73-fa85ed4c905d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.397070] env[62952]: DEBUG oslo_vmware.api [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1023.397070] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b6df5c-8089-91b8-16cc-75cc77193343" [ 1023.397070] env[62952]: _type = "Task" [ 1023.397070] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.397653] env[62952]: DEBUG oslo_vmware.api [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263673, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.327382} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.398040] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1023.398996] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1023.398996] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1023.398996] env[62952]: INFO nova.compute.manager [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1023.398996] env[62952]: DEBUG oslo.service.loopingcall [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.401933] env[62952]: DEBUG nova.compute.manager [-] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1023.402048] env[62952]: DEBUG nova.network.neutron [-] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1023.410619] env[62952]: DEBUG oslo_vmware.api [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b6df5c-8089-91b8-16cc-75cc77193343, 'name': SearchDatastore_Task, 'duration_secs': 0.016964} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.410959] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.411208] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.499330] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263674, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.507495] env[62952]: DEBUG nova.network.neutron [-] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.510095] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263675, 'name': Rename_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.584997] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263676, 'name': Rename_Task, 'duration_secs': 0.307454} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.585440] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1023.585998] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-68b81c66-2504-4a5a-906e-5fab0c19efe3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.594634] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 1023.594634] env[62952]: value = "task-1263677" [ 1023.594634] env[62952]: _type = "Task" [ 1023.594634] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.605135] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263677, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.606415] env[62952]: DEBUG nova.network.neutron [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1023.657316] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263670, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.822346] env[62952]: DEBUG oslo_concurrency.lockutils [req-b0370da5-4189-442b-9cd8-6cede235b97b req-1c4d7d98-8e3b-47cc-8855-ee9236b11dc7 service nova] Releasing lock "refresh_cache-d98a88aa-be5f-4254-aaba-c6fa03c9269a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.014495] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263675, 'name': Rename_Task, 'duration_secs': 0.856251} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.014955] env[62952]: INFO nova.compute.manager [-] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Took 1.54 seconds to deallocate network for instance. [ 1024.015296] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263674, 'name': CreateVM_Task, 'duration_secs': 0.613297} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.018623] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1024.024023] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1024.024608] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3b887949-3582-4a97-ba79-d8bc178856e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.027125] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.027362] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.035026] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1024.035026] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae1294fe-60cb-48f5-9d21-4fe12775dc21 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.047929] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1024.047929] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5293dbda-2991-fa1e-7f77-897771609f22" [ 1024.047929] env[62952]: _type = "Task" [ 1024.047929] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.048349] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Waiting for the task: (returnval){ [ 1024.048349] env[62952]: value = "task-1263678" [ 1024.048349] env[62952]: _type = "Task" [ 1024.048349] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.063216] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5293dbda-2991-fa1e-7f77-897771609f22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.068302] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263678, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.111782] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263677, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.156988] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263670, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.239028] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19a4a5e-f0a2-44d2-b7e0-6dd6fe7cf6d4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.246432] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ad3d1e-6fbb-4623-959d-444d3b2937f5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.281521] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f62002a-6891-45b0-acb7-6e97165bd4a3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.290675] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e70bc7d-422b-4231-8f41-5decae08761f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.306679] env[62952]: DEBUG nova.compute.provider_tree [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.312526] env[62952]: DEBUG nova.compute.manager [req-5f68fbf5-82f8-4a77-b324-2b052cfa9cb0 req-671b7294-670c-46dd-a6af-fd97d2be666a service nova] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Received event network-vif-deleted-5242c318-dbec-47c3-b997-83b63e220ed4 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1024.322644] env[62952]: DEBUG nova.compute.manager [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1024.355316] env[62952]: DEBUG nova.virt.hardware [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1024.355805] env[62952]: DEBUG nova.virt.hardware [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1024.356124] env[62952]: DEBUG nova.virt.hardware [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1024.356518] env[62952]: DEBUG nova.virt.hardware [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1024.356788] env[62952]: DEBUG nova.virt.hardware [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1024.357119] env[62952]: DEBUG nova.virt.hardware [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1024.357503] env[62952]: DEBUG nova.virt.hardware [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1024.357847] env[62952]: DEBUG nova.virt.hardware [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1024.358217] env[62952]: DEBUG nova.virt.hardware [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1024.358544] env[62952]: DEBUG nova.virt.hardware [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1024.358891] env[62952]: DEBUG nova.virt.hardware [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1024.361034] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e363dce-59a9-4098-a352-eca030daa11d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.364623] env[62952]: DEBUG nova.network.neutron [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Updating instance_info_cache with network_info: [{"id": "272ccc8b-6da5-4832-9304-bcf2814b9283", "address": "fa:16:3e:74:e6:82", "network": {"id": "6c5ac793-e327-453e-aff7-69b469be3b76", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1937519631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e340d25add224233aec18cf4c7c62ac3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap272ccc8b-6d", "ovs_interfaceid": "272ccc8b-6da5-4832-9304-bcf2814b9283", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.375027] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95471d73-bd14-48fc-8b72-71b148b6ca92 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.533467] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.565613] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263678, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.566144] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5293dbda-2991-fa1e-7f77-897771609f22, 'name': SearchDatastore_Task, 'duration_secs': 0.060587} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.566551] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.567126] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1024.567485] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.567776] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.568166] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1024.568540] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fd57ce73-2122-4695-a014-77eb7cecd9cd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.580046] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1024.582656] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1024.582656] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0425c038-bb01-43db-bf66-c84f9b951f58 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.588976] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1024.588976] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a8bac9-a490-ac0f-53ca-29447c69f415" [ 1024.588976] env[62952]: _type = "Task" [ 1024.588976] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.599654] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a8bac9-a490-ac0f-53ca-29447c69f415, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.614762] env[62952]: DEBUG oslo_vmware.api [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263677, 'name': PowerOnVM_Task, 'duration_secs': 0.577016} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.615151] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1024.615411] env[62952]: INFO nova.compute.manager [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Took 12.37 seconds to spawn the instance on the hypervisor. [ 1024.615598] env[62952]: DEBUG nova.compute.manager [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1024.616473] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de0da56c-506b-4023-b1b3-4600b1bcdda5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.657227] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263670, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.526295} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.657612] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 1941365e-d562-4d99-89eb-226a07e52071/1941365e-d562-4d99-89eb-226a07e52071.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1024.658126] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1024.658444] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-193195d7-d466-4385-a565-c1cc39819fcc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.667187] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1024.667187] env[62952]: value = "task-1263679" [ 1024.667187] env[62952]: _type = "Task" [ 1024.667187] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.677687] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263679, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.810883] env[62952]: DEBUG nova.network.neutron [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Successfully updated port: fe560683-a241-423f-9f16-c5f59668d967 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1024.810883] env[62952]: DEBUG nova.scheduler.client.report [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1024.867682] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Releasing lock "refresh_cache-4d83faf9-43db-4fc4-978b-55ca9a6ed4c9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.868107] env[62952]: DEBUG nova.compute.manager [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Instance network_info: |[{"id": "272ccc8b-6da5-4832-9304-bcf2814b9283", "address": "fa:16:3e:74:e6:82", "network": {"id": "6c5ac793-e327-453e-aff7-69b469be3b76", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1937519631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e340d25add224233aec18cf4c7c62ac3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap272ccc8b-6d", "ovs_interfaceid": "272ccc8b-6da5-4832-9304-bcf2814b9283", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1024.868550] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:e6:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e044cfd4-1b0d-4d88-b1bd-604025731d3f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '272ccc8b-6da5-4832-9304-bcf2814b9283', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1024.880588] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Creating folder: Project (e340d25add224233aec18cf4c7c62ac3). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1024.881998] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c7b8f73e-3a52-4861-b96c-8fbc87d7586f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.895563] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Created folder: Project (e340d25add224233aec18cf4c7c62ac3) in parent group-v271811. [ 1024.895840] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Creating folder: Instances. Parent ref: group-v271951. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1024.896159] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f423280f-6bd0-4a8f-809a-c11776d57d96 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.909184] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Created folder: Instances in parent group-v271951. [ 1024.912109] env[62952]: DEBUG oslo.service.loopingcall [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1024.912109] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1024.912109] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-404d8fb8-3142-4c88-af97-53bb70da1e5c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.933974] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1024.933974] env[62952]: value = "task-1263682" [ 1024.933974] env[62952]: _type = "Task" [ 1024.933974] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.936615] env[62952]: DEBUG nova.compute.manager [req-6665f505-566c-48a6-913f-571e17244499 req-4505b776-ae51-4053-8740-dbf1c0d2a598 service nova] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Received event network-changed-272ccc8b-6da5-4832-9304-bcf2814b9283 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1024.936853] env[62952]: DEBUG nova.compute.manager [req-6665f505-566c-48a6-913f-571e17244499 req-4505b776-ae51-4053-8740-dbf1c0d2a598 service nova] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Refreshing instance network info cache due to event network-changed-272ccc8b-6da5-4832-9304-bcf2814b9283. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1024.937134] env[62952]: DEBUG oslo_concurrency.lockutils [req-6665f505-566c-48a6-913f-571e17244499 req-4505b776-ae51-4053-8740-dbf1c0d2a598 service nova] Acquiring lock "refresh_cache-4d83faf9-43db-4fc4-978b-55ca9a6ed4c9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.937323] env[62952]: DEBUG oslo_concurrency.lockutils [req-6665f505-566c-48a6-913f-571e17244499 req-4505b776-ae51-4053-8740-dbf1c0d2a598 service nova] Acquired lock "refresh_cache-4d83faf9-43db-4fc4-978b-55ca9a6ed4c9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.937526] env[62952]: DEBUG nova.network.neutron [req-6665f505-566c-48a6-913f-571e17244499 req-4505b776-ae51-4053-8740-dbf1c0d2a598 service nova] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Refreshing network info cache for port 272ccc8b-6da5-4832-9304-bcf2814b9283 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1024.951505] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263682, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.063328] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263678, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.104901] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a8bac9-a490-ac0f-53ca-29447c69f415, 'name': SearchDatastore_Task, 'duration_secs': 0.027695} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.105854] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-599e61c2-a21d-4620-98ef-78c1e087952c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.113732] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1025.113732] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]528074f0-dfc5-e8ee-94ad-05cc488e6363" [ 1025.113732] env[62952]: _type = "Task" [ 1025.113732] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.124591] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]528074f0-dfc5-e8ee-94ad-05cc488e6363, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.142100] env[62952]: INFO nova.compute.manager [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Took 25.86 seconds to build instance. [ 1025.178696] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263679, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.462383} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.178851] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1025.179797] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a8d8fb3-ddc3-471b-a68d-77fa6c8587d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.205473] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 1941365e-d562-4d99-89eb-226a07e52071/1941365e-d562-4d99-89eb-226a07e52071.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1025.206478] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ea31108-aa4a-4abe-a659-3c2bedbe3b43 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.229637] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1025.229637] env[62952]: value = "task-1263683" [ 1025.229637] env[62952]: _type = "Task" [ 1025.229637] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.239497] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263683, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.273566] env[62952]: DEBUG nova.network.neutron [-] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.321415] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "refresh_cache-05223053-90c3-433c-99b9-e252ef9a1935" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.321415] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "refresh_cache-05223053-90c3-433c-99b9-e252ef9a1935" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.321415] env[62952]: DEBUG nova.network.neutron [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1025.451834] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263682, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.562470] env[62952]: DEBUG oslo_vmware.api [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263678, 'name': PowerOnVM_Task, 'duration_secs': 1.052818} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.564856] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1025.565079] env[62952]: INFO nova.compute.manager [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Took 10.89 seconds to spawn the instance on the hypervisor. [ 1025.565268] env[62952]: DEBUG nova.compute.manager [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1025.566155] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70ec114-6927-4374-83fc-02c5c9c8dc9c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.627728] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]528074f0-dfc5-e8ee-94ad-05cc488e6363, 'name': SearchDatastore_Task, 'duration_secs': 0.014646} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.628438] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.628721] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] d98a88aa-be5f-4254-aaba-c6fa03c9269a/d98a88aa-be5f-4254-aaba-c6fa03c9269a.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1025.629071] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5ba685dc-fc48-4963-b3aa-9e3729480bd0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.642205] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1025.642205] env[62952]: value = "task-1263684" [ 1025.642205] env[62952]: _type = "Task" [ 1025.642205] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.647785] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5f887ff-739a-42a1-8ab2-41899ba47fbc tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "120c6804-5aab-4406-9c0c-46eaa835b59f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.378s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.651733] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263684, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.677196] env[62952]: DEBUG nova.network.neutron [req-6665f505-566c-48a6-913f-571e17244499 req-4505b776-ae51-4053-8740-dbf1c0d2a598 service nova] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Updated VIF entry in instance network info cache for port 272ccc8b-6da5-4832-9304-bcf2814b9283. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1025.677467] env[62952]: DEBUG nova.network.neutron [req-6665f505-566c-48a6-913f-571e17244499 req-4505b776-ae51-4053-8740-dbf1c0d2a598 service nova] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Updating instance_info_cache with network_info: [{"id": "272ccc8b-6da5-4832-9304-bcf2814b9283", "address": "fa:16:3e:74:e6:82", "network": {"id": "6c5ac793-e327-453e-aff7-69b469be3b76", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1937519631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e340d25add224233aec18cf4c7c62ac3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap272ccc8b-6d", "ovs_interfaceid": "272ccc8b-6da5-4832-9304-bcf2814b9283", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.741879] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263683, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.757282] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquiring lock "120c6804-5aab-4406-9c0c-46eaa835b59f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.757543] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "120c6804-5aab-4406-9c0c-46eaa835b59f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.757801] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquiring lock "120c6804-5aab-4406-9c0c-46eaa835b59f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.758009] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "120c6804-5aab-4406-9c0c-46eaa835b59f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.758199] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "120c6804-5aab-4406-9c0c-46eaa835b59f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.760746] env[62952]: INFO nova.compute.manager [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Terminating instance [ 1025.762829] env[62952]: DEBUG nova.compute.manager [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1025.763048] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1025.763919] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce100c3-8eb5-4121-a83f-b59e76b8cbb1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.773018] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1025.773430] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f86a21e-eb43-46bc-8880-64eb81f85cb0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.776101] env[62952]: INFO nova.compute.manager [-] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Took 2.37 seconds to deallocate network for instance. [ 1025.785901] env[62952]: DEBUG oslo_vmware.api [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 1025.785901] env[62952]: value = "task-1263685" [ 1025.785901] env[62952]: _type = "Task" [ 1025.785901] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.797455] env[62952]: DEBUG oslo_vmware.api [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263685, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.832753] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.421s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.835633] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.302s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.835763] env[62952]: DEBUG nova.objects.instance [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lazy-loading 'resources' on Instance uuid afaecd3c-da24-4111-8262-d92a57a42a55 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.873265] env[62952]: DEBUG nova.network.neutron [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1025.948166] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263682, 'name': CreateVM_Task, 'duration_secs': 0.5317} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.948166] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1025.948705] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.948868] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.949225] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1025.949510] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00dc7257-85fe-4d70-b527-4c07caeecf6c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.956202] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Waiting for the task: (returnval){ [ 1025.956202] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524fcdb5-0671-20f2-f870-40abffce58d2" [ 1025.956202] env[62952]: _type = "Task" [ 1025.956202] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.966265] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524fcdb5-0671-20f2-f870-40abffce58d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.060933] env[62952]: DEBUG nova.network.neutron [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Updating instance_info_cache with network_info: [{"id": "fe560683-a241-423f-9f16-c5f59668d967", "address": "fa:16:3e:30:43:44", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe560683-a2", "ovs_interfaceid": "fe560683-a241-423f-9f16-c5f59668d967", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.083090] env[62952]: INFO nova.compute.manager [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Took 26.65 seconds to build instance. [ 1026.154376] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263684, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.181827] env[62952]: DEBUG oslo_concurrency.lockutils [req-6665f505-566c-48a6-913f-571e17244499 req-4505b776-ae51-4053-8740-dbf1c0d2a598 service nova] Releasing lock "refresh_cache-4d83faf9-43db-4fc4-978b-55ca9a6ed4c9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.243999] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263683, 'name': ReconfigVM_Task, 'duration_secs': 0.69105} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.244357] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 1941365e-d562-4d99-89eb-226a07e52071/1941365e-d562-4d99-89eb-226a07e52071.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1026.245272] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-55ee89ed-9222-498b-9c3f-b35309d82c86 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.256019] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1026.256019] env[62952]: value = "task-1263686" [ 1026.256019] env[62952]: _type = "Task" [ 1026.256019] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.267774] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263686, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.288350] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.298782] env[62952]: DEBUG oslo_vmware.api [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263685, 'name': PowerOffVM_Task, 'duration_secs': 0.27304} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.299109] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1026.299275] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1026.299539] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c938d3b-bbd9-47c2-8a37-409575139b72 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.348469] env[62952]: DEBUG nova.compute.manager [req-0ab40357-f2b9-4026-a75f-f3f385eeeb89 req-d08308c3-cf8e-48de-b21e-d9275ab2adae service nova] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Received event network-vif-deleted-3a705583-6448-4020-9894-5f2056e2e4e7 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1026.405923] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1026.406256] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1026.406470] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Deleting the datastore file [datastore1] 120c6804-5aab-4406-9c0c-46eaa835b59f {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1026.407587] env[62952]: INFO nova.scheduler.client.report [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleted allocation for migration 51bfa6dd-bc3f-4b5c-89c3-d0da48e18b27 [ 1026.408906] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-788979bd-ead2-4349-9776-30e94d0cf0bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.425635] env[62952]: DEBUG oslo_vmware.api [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for the task: (returnval){ [ 1026.425635] env[62952]: value = "task-1263688" [ 1026.425635] env[62952]: _type = "Task" [ 1026.425635] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.439036] env[62952]: DEBUG oslo_vmware.api [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263688, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.471350] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524fcdb5-0671-20f2-f870-40abffce58d2, 'name': SearchDatastore_Task, 'duration_secs': 0.017674} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.474248] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.474507] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1026.474748] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.474913] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.475116] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1026.476023] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8b333bb7-0f90-4077-96cc-454a83806de3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.498992] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1026.499277] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1026.500216] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b3e0ee1-8545-48d1-9512-2a64ea5dd6ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.511709] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Waiting for the task: (returnval){ [ 1026.511709] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d026c5-bb72-0a18-0a17-46b8686d406b" [ 1026.511709] env[62952]: _type = "Task" [ 1026.511709] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.523872] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d026c5-bb72-0a18-0a17-46b8686d406b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.563460] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "refresh_cache-05223053-90c3-433c-99b9-e252ef9a1935" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.563861] env[62952]: DEBUG nova.compute.manager [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Instance network_info: |[{"id": "fe560683-a241-423f-9f16-c5f59668d967", "address": "fa:16:3e:30:43:44", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe560683-a2", "ovs_interfaceid": "fe560683-a241-423f-9f16-c5f59668d967", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1026.565046] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:43:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e1c9bb98-73a9-48eb-856e-a541afe9b07b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fe560683-a241-423f-9f16-c5f59668d967', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1026.574427] env[62952]: DEBUG oslo.service.loopingcall [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.574584] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1026.574709] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bae18439-9783-4d0a-9e95-1acc0624824d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.591324] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1aef8582-199b-46ba-9100-c54750be6d28 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Lock "511b995d-c531-41f5-9d1b-89ec8c1bfc18" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.167s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.600425] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1026.600425] env[62952]: value = "task-1263689" [ 1026.600425] env[62952]: _type = "Task" [ 1026.600425] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.609224] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263689, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.659315] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263684, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.7094} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.659315] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] d98a88aa-be5f-4254-aaba-c6fa03c9269a/d98a88aa-be5f-4254-aaba-c6fa03c9269a.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1026.659315] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1026.659681] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3cf0cb4d-a17c-41b4-b136-eaaafe3932a2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.671439] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1026.671439] env[62952]: value = "task-1263690" [ 1026.671439] env[62952]: _type = "Task" [ 1026.671439] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.686370] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263690, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.719618] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2460f7e-1a0b-4ad5-8cc3-9323fd0454ea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.728796] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13c0e782-ee8b-4610-8791-fa509b3cd9c0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.767996] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-995e82eb-3ab6-4e26-ac87-5ed8ee677c06 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.777130] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263686, 'name': Rename_Task, 'duration_secs': 0.314373} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.779641] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1026.780028] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b803350f-314a-4ee0-bf53-8ecf4a95e028 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.782856] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1794bf7-adb8-43c1-af3f-ccff2f91e203 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.800611] env[62952]: DEBUG nova.compute.provider_tree [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.803519] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1026.803519] env[62952]: value = "task-1263691" [ 1026.803519] env[62952]: _type = "Task" [ 1026.803519] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.813849] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263691, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.919032] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c91f2fdf-fd60-485a-92da-695b474fd041 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.729s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.939748] env[62952]: DEBUG oslo_vmware.api [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Task: {'id': task-1263688, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.308143} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.941234] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1026.941449] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1026.941648] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1026.941832] env[62952]: INFO nova.compute.manager [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1026.942096] env[62952]: DEBUG oslo.service.loopingcall [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.942618] env[62952]: DEBUG nova.compute.manager [-] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1026.942727] env[62952]: DEBUG nova.network.neutron [-] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1027.023746] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d026c5-bb72-0a18-0a17-46b8686d406b, 'name': SearchDatastore_Task, 'duration_secs': 0.012779} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.025597] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-beffec5b-49d6-4d66-b9d2-d5e9c19fac45 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.030875] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Waiting for the task: (returnval){ [ 1027.030875] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52feec07-2285-782b-1395-984a5d88ecdb" [ 1027.030875] env[62952]: _type = "Task" [ 1027.030875] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.040403] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52feec07-2285-782b-1395-984a5d88ecdb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.087242] env[62952]: DEBUG nova.compute.manager [req-b6271e80-8fda-4347-a100-79096a604422 req-b70a8141-a2f1-48d7-99f3-76f6bec8b5be service nova] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Received event network-vif-plugged-fe560683-a241-423f-9f16-c5f59668d967 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1027.087354] env[62952]: DEBUG oslo_concurrency.lockutils [req-b6271e80-8fda-4347-a100-79096a604422 req-b70a8141-a2f1-48d7-99f3-76f6bec8b5be service nova] Acquiring lock "05223053-90c3-433c-99b9-e252ef9a1935-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.087709] env[62952]: DEBUG oslo_concurrency.lockutils [req-b6271e80-8fda-4347-a100-79096a604422 req-b70a8141-a2f1-48d7-99f3-76f6bec8b5be service nova] Lock "05223053-90c3-433c-99b9-e252ef9a1935-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.088479] env[62952]: DEBUG oslo_concurrency.lockutils [req-b6271e80-8fda-4347-a100-79096a604422 req-b70a8141-a2f1-48d7-99f3-76f6bec8b5be service nova] Lock "05223053-90c3-433c-99b9-e252ef9a1935-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.088857] env[62952]: DEBUG nova.compute.manager [req-b6271e80-8fda-4347-a100-79096a604422 req-b70a8141-a2f1-48d7-99f3-76f6bec8b5be service nova] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] No waiting events found dispatching network-vif-plugged-fe560683-a241-423f-9f16-c5f59668d967 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1027.089196] env[62952]: WARNING nova.compute.manager [req-b6271e80-8fda-4347-a100-79096a604422 req-b70a8141-a2f1-48d7-99f3-76f6bec8b5be service nova] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Received unexpected event network-vif-plugged-fe560683-a241-423f-9f16-c5f59668d967 for instance with vm_state building and task_state spawning. [ 1027.089484] env[62952]: DEBUG nova.compute.manager [req-b6271e80-8fda-4347-a100-79096a604422 req-b70a8141-a2f1-48d7-99f3-76f6bec8b5be service nova] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Received event network-changed-fe560683-a241-423f-9f16-c5f59668d967 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1027.089758] env[62952]: DEBUG nova.compute.manager [req-b6271e80-8fda-4347-a100-79096a604422 req-b70a8141-a2f1-48d7-99f3-76f6bec8b5be service nova] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Refreshing instance network info cache due to event network-changed-fe560683-a241-423f-9f16-c5f59668d967. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1027.090113] env[62952]: DEBUG oslo_concurrency.lockutils [req-b6271e80-8fda-4347-a100-79096a604422 req-b70a8141-a2f1-48d7-99f3-76f6bec8b5be service nova] Acquiring lock "refresh_cache-05223053-90c3-433c-99b9-e252ef9a1935" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.090434] env[62952]: DEBUG oslo_concurrency.lockutils [req-b6271e80-8fda-4347-a100-79096a604422 req-b70a8141-a2f1-48d7-99f3-76f6bec8b5be service nova] Acquired lock "refresh_cache-05223053-90c3-433c-99b9-e252ef9a1935" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.090722] env[62952]: DEBUG nova.network.neutron [req-b6271e80-8fda-4347-a100-79096a604422 req-b70a8141-a2f1-48d7-99f3-76f6bec8b5be service nova] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Refreshing network info cache for port fe560683-a241-423f-9f16-c5f59668d967 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1027.117084] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263689, 'name': CreateVM_Task, 'duration_secs': 0.375839} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.117084] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1027.117084] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.117084] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.117084] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1027.118371] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd3e9680-4669-475b-b509-f1fcc50fbb52 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.126364] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1027.126364] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52823c12-4e0a-1ee1-fd29-80eefafaf58f" [ 1027.126364] env[62952]: _type = "Task" [ 1027.126364] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.138098] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52823c12-4e0a-1ee1-fd29-80eefafaf58f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.188396] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263690, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.123336} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.189107] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1027.190379] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330b146f-cc78-4101-9077-d42d0560b252 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.218024] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] d98a88aa-be5f-4254-aaba-c6fa03c9269a/d98a88aa-be5f-4254-aaba-c6fa03c9269a.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1027.218552] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8609a5a-d5c5-4c9a-8629-ce3afcc9a876 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.243424] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1027.243424] env[62952]: value = "task-1263693" [ 1027.243424] env[62952]: _type = "Task" [ 1027.243424] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.253549] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263693, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.306771] env[62952]: DEBUG nova.scheduler.client.report [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1027.321415] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263691, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.502890] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527ba0cc-0586-d899-029c-860737b22b35/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1027.504656] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dba27ce-6533-401b-8f96-8e840ef7cf23 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.515018] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527ba0cc-0586-d899-029c-860737b22b35/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1027.515018] env[62952]: ERROR oslo_vmware.rw_handles [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527ba0cc-0586-d899-029c-860737b22b35/disk-0.vmdk due to incomplete transfer. [ 1027.515018] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-99fbe8d0-68b2-43db-939f-da60c6139ef3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.524016] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527ba0cc-0586-d899-029c-860737b22b35/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1027.524016] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Uploaded image 13c9a867-a743-44b0-87b1-abaa7ed74f6c to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1027.526947] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1027.527280] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-bbd5e6e5-d17d-47e9-b452-d2d28147b327 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.539471] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1027.539471] env[62952]: value = "task-1263694" [ 1027.539471] env[62952]: _type = "Task" [ 1027.539471] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.548893] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52feec07-2285-782b-1395-984a5d88ecdb, 'name': SearchDatastore_Task, 'duration_secs': 0.020105} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.549725] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.550075] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9/4d83faf9-43db-4fc4-978b-55ca9a6ed4c9.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1027.550397] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1a8d25a-7af7-4399-81f5-7b5a36644fcf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.555818] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263694, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.561424] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Waiting for the task: (returnval){ [ 1027.561424] env[62952]: value = "task-1263695" [ 1027.561424] env[62952]: _type = "Task" [ 1027.561424] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.570410] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': task-1263695, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.641443] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52823c12-4e0a-1ee1-fd29-80eefafaf58f, 'name': SearchDatastore_Task, 'duration_secs': 0.016253} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.641782] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.642037] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1027.642286] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.642439] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.642629] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1027.642910] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bfbc612e-b0a0-4929-b920-65bfa05a6d5e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.654028] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1027.654028] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1027.654028] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b41861d-55cf-46c4-8251-60e74697eef0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.662136] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1027.662136] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e6304f-c755-88a5-8c73-d80c008a0350" [ 1027.662136] env[62952]: _type = "Task" [ 1027.662136] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.676221] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e6304f-c755-88a5-8c73-d80c008a0350, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.758565] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263693, 'name': ReconfigVM_Task, 'duration_secs': 0.317262} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.758927] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Reconfigured VM instance instance-0000005a to attach disk [datastore2] d98a88aa-be5f-4254-aaba-c6fa03c9269a/d98a88aa-be5f-4254-aaba-c6fa03c9269a.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1027.759611] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-250aeef1-1a1e-4383-96ef-4851c1f16f59 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.768311] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1027.768311] env[62952]: value = "task-1263696" [ 1027.768311] env[62952]: _type = "Task" [ 1027.768311] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.778338] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Acquiring lock "511b995d-c531-41f5-9d1b-89ec8c1bfc18" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.778635] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Lock "511b995d-c531-41f5-9d1b-89ec8c1bfc18" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.778807] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Acquiring lock "511b995d-c531-41f5-9d1b-89ec8c1bfc18-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.779073] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Lock "511b995d-c531-41f5-9d1b-89ec8c1bfc18-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.779302] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Lock "511b995d-c531-41f5-9d1b-89ec8c1bfc18-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.781600] env[62952]: INFO nova.compute.manager [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Terminating instance [ 1027.783361] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263696, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.784064] env[62952]: DEBUG nova.compute.manager [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1027.784175] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1027.785010] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de1bbac-1ff0-4065-8bb6-d556c6c03cea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.796827] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1027.797145] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47d90459-c3cb-4e86-9a5f-08892a0310ff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.805588] env[62952]: DEBUG oslo_vmware.api [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Waiting for the task: (returnval){ [ 1027.805588] env[62952]: value = "task-1263697" [ 1027.805588] env[62952]: _type = "Task" [ 1027.805588] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.814602] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.979s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.821741] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.534s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.822186] env[62952]: DEBUG nova.objects.instance [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lazy-loading 'resources' on Instance uuid 9cc7bc2f-ff22-43cf-be79-ff8c6426115b {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.823377] env[62952]: DEBUG oslo_vmware.api [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263697, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.828072] env[62952]: DEBUG oslo_vmware.api [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263691, 'name': PowerOnVM_Task, 'duration_secs': 0.983994} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.828638] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1027.828862] env[62952]: INFO nova.compute.manager [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Took 10.75 seconds to spawn the instance on the hypervisor. [ 1027.829070] env[62952]: DEBUG nova.compute.manager [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1027.830384] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6839fb7d-6bc7-4513-8801-5db8c4e9ace8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.846021] env[62952]: DEBUG nova.network.neutron [req-b6271e80-8fda-4347-a100-79096a604422 req-b70a8141-a2f1-48d7-99f3-76f6bec8b5be service nova] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Updated VIF entry in instance network info cache for port fe560683-a241-423f-9f16-c5f59668d967. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1027.846021] env[62952]: DEBUG nova.network.neutron [req-b6271e80-8fda-4347-a100-79096a604422 req-b70a8141-a2f1-48d7-99f3-76f6bec8b5be service nova] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Updating instance_info_cache with network_info: [{"id": "fe560683-a241-423f-9f16-c5f59668d967", "address": "fa:16:3e:30:43:44", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe560683-a2", "ovs_interfaceid": "fe560683-a241-423f-9f16-c5f59668d967", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.859351] env[62952]: INFO nova.scheduler.client.report [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleted allocations for instance afaecd3c-da24-4111-8262-d92a57a42a55 [ 1028.038967] env[62952]: DEBUG nova.network.neutron [-] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.055790] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263694, 'name': Destroy_Task} progress is 33%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.074886] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': task-1263695, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.173495] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e6304f-c755-88a5-8c73-d80c008a0350, 'name': SearchDatastore_Task, 'duration_secs': 0.015496} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.176018] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3324b06-6031-4341-beff-e319183d4b98 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.183019] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1028.183019] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523e5285-8bd1-d6ee-55e3-884ef440cee2" [ 1028.183019] env[62952]: _type = "Task" [ 1028.183019] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.190328] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523e5285-8bd1-d6ee-55e3-884ef440cee2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.279675] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263696, 'name': Rename_Task, 'duration_secs': 0.183784} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.280287] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1028.280690] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aed8ac9f-0dc9-4b14-af68-6ceb2b7c2259 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.288253] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1028.288253] env[62952]: value = "task-1263698" [ 1028.288253] env[62952]: _type = "Task" [ 1028.288253] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.297304] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263698, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.315985] env[62952]: DEBUG oslo_vmware.api [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263697, 'name': PowerOffVM_Task, 'duration_secs': 0.302611} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.317048] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1028.317048] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1028.317314] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-34de8549-d6cb-4053-bc1a-f736ba0597c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.352226] env[62952]: INFO nova.compute.manager [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Took 26.69 seconds to build instance. [ 1028.355792] env[62952]: DEBUG oslo_concurrency.lockutils [req-b6271e80-8fda-4347-a100-79096a604422 req-b70a8141-a2f1-48d7-99f3-76f6bec8b5be service nova] Releasing lock "refresh_cache-05223053-90c3-433c-99b9-e252ef9a1935" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.367305] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b81c5f4-c2fa-43e8-b0ff-e0b7a2fa3c7c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "afaecd3c-da24-4111-8262-d92a57a42a55" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.965s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.389702] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1028.389952] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1028.390371] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Deleting the datastore file [datastore1] 511b995d-c531-41f5-9d1b-89ec8c1bfc18 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1028.390481] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8945462c-fb74-4777-82c3-a26f3cd03613 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.403513] env[62952]: DEBUG oslo_vmware.api [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Waiting for the task: (returnval){ [ 1028.403513] env[62952]: value = "task-1263700" [ 1028.403513] env[62952]: _type = "Task" [ 1028.403513] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.412574] env[62952]: DEBUG oslo_vmware.api [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263700, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.540783] env[62952]: INFO nova.compute.manager [-] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Took 1.60 seconds to deallocate network for instance. [ 1028.550709] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-636e249a-df62-4007-b8ef-583cf905c0c1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.556267] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263694, 'name': Destroy_Task, 'duration_secs': 0.868783} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.556962] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Destroyed the VM [ 1028.557364] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1028.557615] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d81b4527-8561-4a97-bbd8-2f0aa0c3da09 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.562832] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-883818aa-bb74-4526-a69f-d5be77575297 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.571185] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1028.571185] env[62952]: value = "task-1263701" [ 1028.571185] env[62952]: _type = "Task" [ 1028.571185] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.603595] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': task-1263695, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.631472} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.605223] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9/4d83faf9-43db-4fc4-978b-55ca9a6ed4c9.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1028.605451] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1028.606231] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41295f3e-5de2-465e-ac47-cce15b1cfd33 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.612269] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-adf846af-73c1-472b-a4d6-130e05417163 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.614452] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263701, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.626332] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c48d827-fd0b-45e1-aa83-3d8d27e325b1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.631208] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Waiting for the task: (returnval){ [ 1028.631208] env[62952]: value = "task-1263702" [ 1028.631208] env[62952]: _type = "Task" [ 1028.631208] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.649403] env[62952]: DEBUG nova.compute.provider_tree [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.657241] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': task-1263702, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.692642] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523e5285-8bd1-d6ee-55e3-884ef440cee2, 'name': SearchDatastore_Task, 'duration_secs': 0.017525} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.692918] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.693205] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 05223053-90c3-433c-99b9-e252ef9a1935/05223053-90c3-433c-99b9-e252ef9a1935.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1028.693475] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aff84010-0817-436f-b597-45e9c7cdc375 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.700993] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1028.700993] env[62952]: value = "task-1263703" [ 1028.700993] env[62952]: _type = "Task" [ 1028.700993] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.709510] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263703, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.799151] env[62952]: DEBUG oslo_vmware.api [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263698, 'name': PowerOnVM_Task, 'duration_secs': 0.50138} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.799570] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1028.799708] env[62952]: INFO nova.compute.manager [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Took 9.28 seconds to spawn the instance on the hypervisor. [ 1028.799957] env[62952]: DEBUG nova.compute.manager [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1028.800892] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04440023-4907-4502-95b4-6c2260b1a09b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.854098] env[62952]: DEBUG oslo_concurrency.lockutils [None req-09f9b2ea-abff-4314-9e31-d997ba32b2fb tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "1941365e-d562-4d99-89eb-226a07e52071" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.219s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.915045] env[62952]: DEBUG oslo_vmware.api [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Task: {'id': task-1263700, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.480926} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.915326] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1028.915520] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1028.915700] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1028.915876] env[62952]: INFO nova.compute.manager [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1028.916144] env[62952]: DEBUG oslo.service.loopingcall [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.917331] env[62952]: DEBUG nova.compute.manager [-] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1028.917661] env[62952]: DEBUG nova.network.neutron [-] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1029.057609] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.082924] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263701, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.124963] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.125368] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.125599] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.125801] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.125974] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.129749] env[62952]: INFO nova.compute.manager [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Terminating instance [ 1029.134283] env[62952]: DEBUG nova.compute.manager [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1029.134697] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1029.135446] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebcd74e0-d3a6-4d9c-a365-355589ee8870 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.153595] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1029.155417] env[62952]: DEBUG nova.scheduler.client.report [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.158895] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': task-1263702, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073623} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.159345] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-02673f37-33b4-410f-ba1c-eeed62143828 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.161429] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1029.165008] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2aa2e7c-e4b4-41cf-b612-3d40a42ab7c0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.188519] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9/4d83faf9-43db-4fc4-978b-55ca9a6ed4c9.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1029.192135] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8289d40a-7267-4719-a334-3dd3a0ca13e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.208951] env[62952]: DEBUG nova.compute.manager [req-d96d8faf-199d-4df5-af01-855e770b39b2 req-c812778c-5e56-4c78-82d8-b32e6c128031 service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Received event network-vif-deleted-a8f57522-7ac2-454d-9bac-0be6f7897f13 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1029.209176] env[62952]: DEBUG nova.compute.manager [req-d96d8faf-199d-4df5-af01-855e770b39b2 req-c812778c-5e56-4c78-82d8-b32e6c128031 service nova] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Received event network-vif-deleted-0a105c0f-99dd-4c52-b3b4-72de6af13865 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1029.210551] env[62952]: DEBUG oslo_vmware.api [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1029.210551] env[62952]: value = "task-1263705" [ 1029.210551] env[62952]: _type = "Task" [ 1029.210551] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.224583] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Waiting for the task: (returnval){ [ 1029.224583] env[62952]: value = "task-1263706" [ 1029.224583] env[62952]: _type = "Task" [ 1029.224583] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.231799] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263703, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.232144] env[62952]: DEBUG oslo_vmware.api [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263705, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.244859] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': task-1263706, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.320364] env[62952]: INFO nova.compute.manager [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Took 24.42 seconds to build instance. [ 1029.423327] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "6f0b04f5-811f-4c53-808b-6d9d22100a86" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.423653] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "6f0b04f5-811f-4c53-808b-6d9d22100a86" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.423894] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "6f0b04f5-811f-4c53-808b-6d9d22100a86-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.424203] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "6f0b04f5-811f-4c53-808b-6d9d22100a86-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.424495] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "6f0b04f5-811f-4c53-808b-6d9d22100a86-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.427697] env[62952]: INFO nova.compute.manager [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Terminating instance [ 1029.430205] env[62952]: DEBUG nova.compute.manager [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1029.430445] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1029.431435] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3410b8-b983-4c6c-86c5-76f4fe04b605 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.441056] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1029.441365] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-36b810c9-937e-4c08-a2b9-8bad8627784f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.451544] env[62952]: DEBUG oslo_vmware.api [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1029.451544] env[62952]: value = "task-1263707" [ 1029.451544] env[62952]: _type = "Task" [ 1029.451544] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.462526] env[62952]: DEBUG oslo_vmware.api [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263707, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.586326] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263701, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.666099] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.844s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.671472] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.671472] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.673049] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.615s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.673480] env[62952]: DEBUG nova.objects.instance [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lazy-loading 'resources' on Instance uuid 120c6804-5aab-4406-9c0c-46eaa835b59f {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.696996] env[62952]: INFO nova.scheduler.client.report [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Deleted allocations for instance 9cc7bc2f-ff22-43cf-be79-ff8c6426115b [ 1029.732031] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263703, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.836047} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.732031] env[62952]: DEBUG oslo_vmware.api [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263705, 'name': PowerOffVM_Task, 'duration_secs': 0.478156} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.732031] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 05223053-90c3-433c-99b9-e252ef9a1935/05223053-90c3-433c-99b9-e252ef9a1935.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1029.732031] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1029.732031] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1029.732031] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1029.732031] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3bf58349-c610-4dc0-b38a-6f30d49c415a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.733316] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-06674bf7-1607-4354-b0f7-a95d9d7291cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.746134] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': task-1263706, 'name': ReconfigVM_Task, 'duration_secs': 0.415009} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.748478] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9/4d83faf9-43db-4fc4-978b-55ca9a6ed4c9.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1029.749218] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1029.749218] env[62952]: value = "task-1263708" [ 1029.749218] env[62952]: _type = "Task" [ 1029.749218] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.749429] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bb610f7d-bca5-4b6f-a3af-f41cf7aff291 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.755155] env[62952]: DEBUG nova.network.neutron [-] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.762509] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263708, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.764370] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Waiting for the task: (returnval){ [ 1029.764370] env[62952]: value = "task-1263710" [ 1029.764370] env[62952]: _type = "Task" [ 1029.764370] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.778469] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': task-1263710, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.824679] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6ae51c4a-259a-4ef4-8fe5-06d80118fed7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "d98a88aa-be5f-4254-aaba-c6fa03c9269a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.932s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.852352] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1029.852817] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1029.853866] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleting the datastore file [datastore2] 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1029.853866] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0cafb587-69af-44ef-b63b-9682813cffc9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.863385] env[62952]: DEBUG oslo_vmware.api [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1029.863385] env[62952]: value = "task-1263711" [ 1029.863385] env[62952]: _type = "Task" [ 1029.863385] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.875544] env[62952]: DEBUG oslo_vmware.api [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263711, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.962610] env[62952]: DEBUG oslo_vmware.api [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263707, 'name': PowerOffVM_Task, 'duration_secs': 0.29014} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.962901] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1029.963091] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1029.963352] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c204d88d-1747-4a6a-80aa-7761d28b20d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.048634] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1030.049022] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1030.049303] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleting the datastore file [datastore2] 6f0b04f5-811f-4c53-808b-6d9d22100a86 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1030.049665] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-213b88f2-1b99-415a-8202-c2d44eb2e59f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.057546] env[62952]: DEBUG oslo_vmware.api [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1030.057546] env[62952]: value = "task-1263713" [ 1030.057546] env[62952]: _type = "Task" [ 1030.057546] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.066183] env[62952]: DEBUG oslo_vmware.api [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263713, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.081277] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263701, 'name': RemoveSnapshot_Task, 'duration_secs': 1.164746} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.081527] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1030.081799] env[62952]: DEBUG nova.compute.manager [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1030.082619] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be307414-a980-4743-9801-7fb391f0d42b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.110574] env[62952]: INFO nova.compute.manager [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Rescuing [ 1030.111095] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "refresh_cache-d98a88aa-be5f-4254-aaba-c6fa03c9269a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.111095] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquired lock "refresh_cache-d98a88aa-be5f-4254-aaba-c6fa03c9269a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.111390] env[62952]: DEBUG nova.network.neutron [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1030.173354] env[62952]: DEBUG nova.compute.manager [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1030.205240] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ff42691-76a9-42bd-ac71-5e4777615940 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "9cc7bc2f-ff22-43cf-be79-ff8c6426115b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.968s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.257984] env[62952]: INFO nova.compute.manager [-] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Took 1.34 seconds to deallocate network for instance. [ 1030.263248] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263708, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.164227} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.268352] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1030.269602] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f21b7e-e878-43e6-b493-54752cd75083 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.280778] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': task-1263710, 'name': Rename_Task, 'duration_secs': 0.262682} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.292202] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1030.301588] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 05223053-90c3-433c-99b9-e252ef9a1935/05223053-90c3-433c-99b9-e252ef9a1935.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1030.303122] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75d787ba-5266-41ae-b83b-260e81249a28 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.304621] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eed587fe-554d-4799-b63e-b677f3744b2a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.333639] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1030.333639] env[62952]: value = "task-1263715" [ 1030.333639] env[62952]: _type = "Task" [ 1030.333639] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.335736] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Waiting for the task: (returnval){ [ 1030.335736] env[62952]: value = "task-1263714" [ 1030.335736] env[62952]: _type = "Task" [ 1030.335736] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.350367] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263715, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.353754] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': task-1263714, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.379373] env[62952]: DEBUG oslo_vmware.api [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263711, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.439964] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-440efcca-b8bc-4eac-8f7f-453aef93be36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.449569] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb042dda-8a9a-4ffc-a684-8f3806ebce45 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.502972] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e73c79-0141-42a9-9cc0-884d18fb431d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.514575] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24341132-07c3-4716-a501-757299f59c95 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.531333] env[62952]: DEBUG nova.compute.provider_tree [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.567371] env[62952]: DEBUG oslo_vmware.api [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263713, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.481164} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.567719] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1030.567919] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1030.568134] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1030.570469] env[62952]: INFO nova.compute.manager [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1030.571082] env[62952]: DEBUG oslo.service.loopingcall [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1030.571212] env[62952]: DEBUG nova.compute.manager [-] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1030.571313] env[62952]: DEBUG nova.network.neutron [-] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1030.595283] env[62952]: INFO nova.compute.manager [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Shelve offloading [ 1030.599489] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1030.599489] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08dba6fe-7ffb-4c42-8c53-f27dfe028433 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.607674] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1030.607674] env[62952]: value = "task-1263716" [ 1030.607674] env[62952]: _type = "Task" [ 1030.607674] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.625564] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1030.629017] env[62952]: DEBUG nova.compute.manager [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1030.629017] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d21b2ba9-3457-4673-a894-7f077c0c2522 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.635182] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "refresh_cache-063d1d92-0526-48e0-b33c-f51a79e7c821" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.635443] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "refresh_cache-063d1d92-0526-48e0-b33c-f51a79e7c821" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.635725] env[62952]: DEBUG nova.network.neutron [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1030.709341] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.770143] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.848200] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263715, 'name': ReconfigVM_Task, 'duration_secs': 0.444329} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.848888] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 05223053-90c3-433c-99b9-e252ef9a1935/05223053-90c3-433c-99b9-e252ef9a1935.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1030.849882] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf3cabf1-9d42-4fa8-9e21-deac784a4fd5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.855960] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': task-1263714, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.863598] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1030.863598] env[62952]: value = "task-1263717" [ 1030.863598] env[62952]: _type = "Task" [ 1030.863598] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.879510] env[62952]: DEBUG oslo_vmware.api [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263711, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.536906} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.883000] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1030.883246] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1030.883515] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1030.883629] env[62952]: INFO nova.compute.manager [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Took 1.75 seconds to destroy the instance on the hypervisor. [ 1030.883897] env[62952]: DEBUG oslo.service.loopingcall [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1030.884197] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263717, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.884685] env[62952]: DEBUG nova.compute.manager [-] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1030.884786] env[62952]: DEBUG nova.network.neutron [-] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1031.040220] env[62952]: DEBUG nova.scheduler.client.report [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1031.045827] env[62952]: DEBUG nova.network.neutron [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Updating instance_info_cache with network_info: [{"id": "3dff9ce5-4a76-49a1-ac02-2702ccde508f", "address": "fa:16:3e:95:0b:fe", "network": {"id": "d84b900e-d793-4187-aaeb-bff9b040e607", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1370737473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a6fddef5a2114acb9749b1bae171cbca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dff9ce5-4a", "ovs_interfaceid": "3dff9ce5-4a76-49a1-ac02-2702ccde508f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.218128] env[62952]: DEBUG nova.compute.manager [req-339335a4-dc8b-4362-983a-b20d4bb9bfb2 req-713947cd-74be-4dc0-a6e3-d4aa514633dd service nova] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Received event network-vif-deleted-a11e8e2c-69be-436f-af5b-48f53b0495c8 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1031.218128] env[62952]: DEBUG nova.compute.manager [req-339335a4-dc8b-4362-983a-b20d4bb9bfb2 req-713947cd-74be-4dc0-a6e3-d4aa514633dd service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Received event network-changed-e1c235af-9501-47c0-a83a-ba40ea7fa8c1 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1031.218128] env[62952]: DEBUG nova.compute.manager [req-339335a4-dc8b-4362-983a-b20d4bb9bfb2 req-713947cd-74be-4dc0-a6e3-d4aa514633dd service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Refreshing instance network info cache due to event network-changed-e1c235af-9501-47c0-a83a-ba40ea7fa8c1. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1031.218442] env[62952]: DEBUG oslo_concurrency.lockutils [req-339335a4-dc8b-4362-983a-b20d4bb9bfb2 req-713947cd-74be-4dc0-a6e3-d4aa514633dd service nova] Acquiring lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.218442] env[62952]: DEBUG oslo_concurrency.lockutils [req-339335a4-dc8b-4362-983a-b20d4bb9bfb2 req-713947cd-74be-4dc0-a6e3-d4aa514633dd service nova] Acquired lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.218641] env[62952]: DEBUG nova.network.neutron [req-339335a4-dc8b-4362-983a-b20d4bb9bfb2 req-713947cd-74be-4dc0-a6e3-d4aa514633dd service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Refreshing network info cache for port e1c235af-9501-47c0-a83a-ba40ea7fa8c1 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1031.356645] env[62952]: DEBUG nova.network.neutron [-] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.356645] env[62952]: DEBUG oslo_vmware.api [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': task-1263714, 'name': PowerOnVM_Task, 'duration_secs': 0.630003} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.356645] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1031.356645] env[62952]: INFO nova.compute.manager [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Took 9.43 seconds to spawn the instance on the hypervisor. [ 1031.356645] env[62952]: DEBUG nova.compute.manager [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1031.356645] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3388d17-44bc-44e3-b675-defb62cfd883 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.378324] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263717, 'name': Rename_Task, 'duration_secs': 0.198933} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.379285] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1031.379621] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-111e4380-eab4-41a7-8667-cbc88f4152df {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.387829] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1031.387829] env[62952]: value = "task-1263719" [ 1031.387829] env[62952]: _type = "Task" [ 1031.387829] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.398252] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263719, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.451888] env[62952]: DEBUG nova.network.neutron [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Updating instance_info_cache with network_info: [{"id": "b0b497d8-449d-43f5-9f1a-ff2172ae47db", "address": "fa:16:3e:f3:4f:7b", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0b497d8-44", "ovs_interfaceid": "b0b497d8-449d-43f5-9f1a-ff2172ae47db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.548075] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.875s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.550551] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.841s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.552344] env[62952]: INFO nova.compute.claims [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1031.556259] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Releasing lock "refresh_cache-d98a88aa-be5f-4254-aaba-c6fa03c9269a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.578988] env[62952]: INFO nova.scheduler.client.report [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Deleted allocations for instance 120c6804-5aab-4406-9c0c-46eaa835b59f [ 1031.653073] env[62952]: DEBUG nova.network.neutron [-] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.856393] env[62952]: INFO nova.compute.manager [-] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Took 1.28 seconds to deallocate network for instance. [ 1031.882147] env[62952]: INFO nova.compute.manager [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Took 23.20 seconds to build instance. [ 1031.904920] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263719, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.955242] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "refresh_cache-063d1d92-0526-48e0-b33c-f51a79e7c821" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.091333] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fa1c9a35-aabd-4ba8-b356-5c65f28ed9d2 tempest-ServersTestMultiNic-667179236 tempest-ServersTestMultiNic-667179236-project-member] Lock "120c6804-5aab-4406-9c0c-46eaa835b59f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.334s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.099955] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1032.100248] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-749c1793-f80a-4050-b1c7-3a61095ce0c1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.114159] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1032.114159] env[62952]: value = "task-1263720" [ 1032.114159] env[62952]: _type = "Task" [ 1032.114159] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.124886] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263720, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.136048] env[62952]: DEBUG nova.network.neutron [req-339335a4-dc8b-4362-983a-b20d4bb9bfb2 req-713947cd-74be-4dc0-a6e3-d4aa514633dd service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Updated VIF entry in instance network info cache for port e1c235af-9501-47c0-a83a-ba40ea7fa8c1. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1032.136442] env[62952]: DEBUG nova.network.neutron [req-339335a4-dc8b-4362-983a-b20d4bb9bfb2 req-713947cd-74be-4dc0-a6e3-d4aa514633dd service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Updating instance_info_cache with network_info: [{"id": "e1c235af-9501-47c0-a83a-ba40ea7fa8c1", "address": "fa:16:3e:34:07:9b", "network": {"id": "827a8171-6759-4434-a6f1-437aad881c81", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1902486970-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091cdeb6da48492bb02f93822a45c9bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11b669be-fb26-4ef8-bdb6-c77ab9d06daf", "external-id": "nsx-vlan-transportzone-633", "segmentation_id": 633, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1c235af-95", "ovs_interfaceid": "e1c235af-9501-47c0-a83a-ba40ea7fa8c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.156587] env[62952]: INFO nova.compute.manager [-] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Took 1.27 seconds to deallocate network for instance. [ 1032.233346] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1032.234282] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e1b962-59ee-4eee-a88c-51fcbd974b33 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.242616] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1032.242920] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b1da7ce-9c7a-4e0f-a796-c73d54640d3e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.367392] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.385286] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a39d8b54-a71c-4484-8a27-954375da38ba tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Lock "4d83faf9-43db-4fc4-978b-55ca9a6ed4c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.715s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.400461] env[62952]: DEBUG oslo_vmware.api [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263719, 'name': PowerOnVM_Task, 'duration_secs': 0.804285} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.400827] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1032.401100] env[62952]: INFO nova.compute.manager [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Took 8.08 seconds to spawn the instance on the hypervisor. [ 1032.401514] env[62952]: DEBUG nova.compute.manager [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1032.403947] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e9c4a5e-ce34-49b4-a91a-fdc96480f576 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.407464] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1032.407821] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1032.408107] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleting the datastore file [datastore2] 063d1d92-0526-48e0-b33c-f51a79e7c821 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1032.408406] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ccb65ae4-1702-42fa-b7c9-6553dd35f543 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.421834] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1032.421834] env[62952]: value = "task-1263722" [ 1032.421834] env[62952]: _type = "Task" [ 1032.421834] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.433873] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263722, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.624675] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263720, 'name': PowerOffVM_Task, 'duration_secs': 0.302768} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.628779] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1032.629500] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d469b1b-abc2-4a56-bc38-b08dce781d2c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.651799] env[62952]: DEBUG oslo_concurrency.lockutils [req-339335a4-dc8b-4362-983a-b20d4bb9bfb2 req-713947cd-74be-4dc0-a6e3-d4aa514633dd service nova] Releasing lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.652171] env[62952]: DEBUG nova.compute.manager [req-339335a4-dc8b-4362-983a-b20d4bb9bfb2 req-713947cd-74be-4dc0-a6e3-d4aa514633dd service nova] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Received event network-vif-deleted-7847ca5f-ad23-49b3-b3e9-6ef833b3c964 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1032.652419] env[62952]: INFO nova.compute.manager [req-339335a4-dc8b-4362-983a-b20d4bb9bfb2 req-713947cd-74be-4dc0-a6e3-d4aa514633dd service nova] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Neutron deleted interface 7847ca5f-ad23-49b3-b3e9-6ef833b3c964; detaching it from the instance and deleting it from the info cache [ 1032.652601] env[62952]: DEBUG nova.network.neutron [req-339335a4-dc8b-4362-983a-b20d4bb9bfb2 req-713947cd-74be-4dc0-a6e3-d4aa514633dd service nova] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.658269] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88cffb24-64ba-4d44-ad18-97fc66df0db7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.668159] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.699559] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1032.699559] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0cfe0f0e-a2ba-4c69-ac3e-ede9ed52e9e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.707033] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1032.707033] env[62952]: value = "task-1263723" [ 1032.707033] env[62952]: _type = "Task" [ 1032.707033] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.719143] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1032.719383] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1032.719820] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.719820] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.720015] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1032.720282] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-555bfe21-3e99-4257-9fe8-14741dfa3ca4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.728955] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1032.729067] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1032.729777] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-763044b6-3cc8-4498-9c96-7b863bb2f418 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.734846] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1032.734846] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52fe3596-9fd5-4c43-092a-78537d1471de" [ 1032.734846] env[62952]: _type = "Task" [ 1032.734846] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.746418] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52fe3596-9fd5-4c43-092a-78537d1471de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.798766] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a0340b-ea07-49b7-a404-37562cddf32d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.807068] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3cbbb0-25d6-422f-9a0f-424a03d7aac5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.837704] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df0d79c6-9a68-475f-b269-daf60606196f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.845615] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9066b91b-cd84-439b-ac43-82aabd3819ac {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.862313] env[62952]: DEBUG nova.compute.provider_tree [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.934150] env[62952]: INFO nova.compute.manager [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Took 23.60 seconds to build instance. [ 1032.938659] env[62952]: DEBUG oslo_vmware.api [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263722, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.431886} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.938897] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1032.939237] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1032.939332] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1033.012648] env[62952]: DEBUG oslo_concurrency.lockutils [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Acquiring lock "4d83faf9-43db-4fc4-978b-55ca9a6ed4c9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.013520] env[62952]: DEBUG oslo_concurrency.lockutils [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Lock "4d83faf9-43db-4fc4-978b-55ca9a6ed4c9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.013520] env[62952]: DEBUG oslo_concurrency.lockutils [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Acquiring lock "4d83faf9-43db-4fc4-978b-55ca9a6ed4c9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.014236] env[62952]: DEBUG oslo_concurrency.lockutils [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Lock "4d83faf9-43db-4fc4-978b-55ca9a6ed4c9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.014236] env[62952]: DEBUG oslo_concurrency.lockutils [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Lock "4d83faf9-43db-4fc4-978b-55ca9a6ed4c9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.017112] env[62952]: INFO nova.compute.manager [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Terminating instance [ 1033.022032] env[62952]: DEBUG nova.compute.manager [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1033.022241] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1033.023255] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ae8749-be2a-42ae-847e-f826fbdb09ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.031046] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1033.031400] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-44f29aa0-0a68-44bf-8893-547a390dba0d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.037535] env[62952]: DEBUG oslo_vmware.api [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Waiting for the task: (returnval){ [ 1033.037535] env[62952]: value = "task-1263724" [ 1033.037535] env[62952]: _type = "Task" [ 1033.037535] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.048802] env[62952]: DEBUG oslo_vmware.api [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': task-1263724, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.050388] env[62952]: INFO nova.scheduler.client.report [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleted allocations for instance 063d1d92-0526-48e0-b33c-f51a79e7c821 [ 1033.158663] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5576fe79-6945-464a-a1f5-dab860e153e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.172520] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed383d9b-b40e-49b0-9695-bf37742919d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.198817] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "42793459-89e6-4941-9b20-cbe52a241c6d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.198817] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "42793459-89e6-4941-9b20-cbe52a241c6d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.222237] env[62952]: DEBUG nova.compute.manager [req-339335a4-dc8b-4362-983a-b20d4bb9bfb2 req-713947cd-74be-4dc0-a6e3-d4aa514633dd service nova] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Detach interface failed, port_id=7847ca5f-ad23-49b3-b3e9-6ef833b3c964, reason: Instance 6f0b04f5-811f-4c53-808b-6d9d22100a86 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1033.222711] env[62952]: DEBUG nova.compute.manager [req-339335a4-dc8b-4362-983a-b20d4bb9bfb2 req-713947cd-74be-4dc0-a6e3-d4aa514633dd service nova] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Received event network-vif-deleted-ea5399c4-f18f-4c0b-82b1-062064db974b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1033.223047] env[62952]: INFO nova.compute.manager [req-339335a4-dc8b-4362-983a-b20d4bb9bfb2 req-713947cd-74be-4dc0-a6e3-d4aa514633dd service nova] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Neutron deleted interface ea5399c4-f18f-4c0b-82b1-062064db974b; detaching it from the instance and deleting it from the info cache [ 1033.223374] env[62952]: DEBUG nova.network.neutron [req-339335a4-dc8b-4362-983a-b20d4bb9bfb2 req-713947cd-74be-4dc0-a6e3-d4aa514633dd service nova] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.248093] env[62952]: DEBUG nova.compute.manager [req-dc5d31ae-55d5-43d1-87b9-07008de73fcc req-a2bc179d-e834-432c-a26c-6f18a8df8125 service nova] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Received event network-vif-unplugged-b0b497d8-449d-43f5-9f1a-ff2172ae47db {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1033.248093] env[62952]: DEBUG oslo_concurrency.lockutils [req-dc5d31ae-55d5-43d1-87b9-07008de73fcc req-a2bc179d-e834-432c-a26c-6f18a8df8125 service nova] Acquiring lock "063d1d92-0526-48e0-b33c-f51a79e7c821-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.248616] env[62952]: DEBUG oslo_concurrency.lockutils [req-dc5d31ae-55d5-43d1-87b9-07008de73fcc req-a2bc179d-e834-432c-a26c-6f18a8df8125 service nova] Lock "063d1d92-0526-48e0-b33c-f51a79e7c821-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.248616] env[62952]: DEBUG oslo_concurrency.lockutils [req-dc5d31ae-55d5-43d1-87b9-07008de73fcc req-a2bc179d-e834-432c-a26c-6f18a8df8125 service nova] Lock "063d1d92-0526-48e0-b33c-f51a79e7c821-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.248706] env[62952]: DEBUG nova.compute.manager [req-dc5d31ae-55d5-43d1-87b9-07008de73fcc req-a2bc179d-e834-432c-a26c-6f18a8df8125 service nova] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] No waiting events found dispatching network-vif-unplugged-b0b497d8-449d-43f5-9f1a-ff2172ae47db {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1033.248844] env[62952]: WARNING nova.compute.manager [req-dc5d31ae-55d5-43d1-87b9-07008de73fcc req-a2bc179d-e834-432c-a26c-6f18a8df8125 service nova] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Received unexpected event network-vif-unplugged-b0b497d8-449d-43f5-9f1a-ff2172ae47db for instance with vm_state shelved_offloaded and task_state None. [ 1033.249047] env[62952]: DEBUG nova.compute.manager [req-dc5d31ae-55d5-43d1-87b9-07008de73fcc req-a2bc179d-e834-432c-a26c-6f18a8df8125 service nova] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Received event network-changed-b0b497d8-449d-43f5-9f1a-ff2172ae47db {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1033.249225] env[62952]: DEBUG nova.compute.manager [req-dc5d31ae-55d5-43d1-87b9-07008de73fcc req-a2bc179d-e834-432c-a26c-6f18a8df8125 service nova] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Refreshing instance network info cache due to event network-changed-b0b497d8-449d-43f5-9f1a-ff2172ae47db. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1033.249429] env[62952]: DEBUG oslo_concurrency.lockutils [req-dc5d31ae-55d5-43d1-87b9-07008de73fcc req-a2bc179d-e834-432c-a26c-6f18a8df8125 service nova] Acquiring lock "refresh_cache-063d1d92-0526-48e0-b33c-f51a79e7c821" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.249609] env[62952]: DEBUG oslo_concurrency.lockutils [req-dc5d31ae-55d5-43d1-87b9-07008de73fcc req-a2bc179d-e834-432c-a26c-6f18a8df8125 service nova] Acquired lock "refresh_cache-063d1d92-0526-48e0-b33c-f51a79e7c821" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.249834] env[62952]: DEBUG nova.network.neutron [req-dc5d31ae-55d5-43d1-87b9-07008de73fcc req-a2bc179d-e834-432c-a26c-6f18a8df8125 service nova] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Refreshing network info cache for port b0b497d8-449d-43f5-9f1a-ff2172ae47db {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1033.255255] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52fe3596-9fd5-4c43-092a-78537d1471de, 'name': SearchDatastore_Task, 'duration_secs': 0.011988} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.256270] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6941756-9a09-4ea6-aec2-918b8f3c45d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.263333] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1033.263333] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526b236e-3ea2-9775-425c-2696b34cd144" [ 1033.263333] env[62952]: _type = "Task" [ 1033.263333] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.272893] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526b236e-3ea2-9775-425c-2696b34cd144, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.367376] env[62952]: DEBUG nova.scheduler.client.report [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1033.439589] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b24b155-91ef-40ba-86e2-2adde3eb702a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "05223053-90c3-433c-99b9-e252ef9a1935" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.116s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.548353] env[62952]: DEBUG oslo_vmware.api [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': task-1263724, 'name': PowerOffVM_Task, 'duration_secs': 0.348175} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.548627] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1033.548799] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1033.549066] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-09e8a82d-cbdc-411a-a1ea-6ad07242a655 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.554159] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.602927] env[62952]: INFO nova.compute.manager [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Rebuilding instance [ 1033.635894] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1033.635894] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1033.636191] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Deleting the datastore file [datastore2] 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1033.636504] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b8f725b-02a3-4f5b-918d-7150ca0733f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.649801] env[62952]: DEBUG oslo_vmware.api [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Waiting for the task: (returnval){ [ 1033.649801] env[62952]: value = "task-1263726" [ 1033.649801] env[62952]: _type = "Task" [ 1033.649801] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.659454] env[62952]: DEBUG oslo_vmware.api [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': task-1263726, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.669031] env[62952]: DEBUG nova.compute.manager [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1033.669031] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd20ddd7-3004-4c3f-b9ce-2872dcc68aec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.700278] env[62952]: DEBUG nova.compute.manager [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1033.731706] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f81fb72-7e72-4056-830a-3cce36bfbd23 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.747963] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-935abfcb-d36e-4233-8fc0-8858d4eeb81d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.773036] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526b236e-3ea2-9775-425c-2696b34cd144, 'name': SearchDatastore_Task, 'duration_secs': 0.019014} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.782397] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.782830] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] d98a88aa-be5f-4254-aaba-c6fa03c9269a/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk. {{(pid=62952) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1033.783589] env[62952]: DEBUG nova.compute.manager [req-339335a4-dc8b-4362-983a-b20d4bb9bfb2 req-713947cd-74be-4dc0-a6e3-d4aa514633dd service nova] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Detach interface failed, port_id=ea5399c4-f18f-4c0b-82b1-062064db974b, reason: Instance 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1033.783982] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ca9818e8-96e5-41d6-875d-8b7768a06159 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.791200] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1033.791200] env[62952]: value = "task-1263727" [ 1033.791200] env[62952]: _type = "Task" [ 1033.791200] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.801766] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263727, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.873026] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.873671] env[62952]: DEBUG nova.compute.manager [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1033.886330] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.116s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.886620] env[62952]: DEBUG nova.objects.instance [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Lazy-loading 'resources' on Instance uuid 511b995d-c531-41f5-9d1b-89ec8c1bfc18 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.006561] env[62952]: DEBUG nova.network.neutron [req-dc5d31ae-55d5-43d1-87b9-07008de73fcc req-a2bc179d-e834-432c-a26c-6f18a8df8125 service nova] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Updated VIF entry in instance network info cache for port b0b497d8-449d-43f5-9f1a-ff2172ae47db. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1034.007091] env[62952]: DEBUG nova.network.neutron [req-dc5d31ae-55d5-43d1-87b9-07008de73fcc req-a2bc179d-e834-432c-a26c-6f18a8df8125 service nova] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Updating instance_info_cache with network_info: [{"id": "b0b497d8-449d-43f5-9f1a-ff2172ae47db", "address": "fa:16:3e:f3:4f:7b", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": null, "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapb0b497d8-44", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.159929] env[62952]: DEBUG oslo_vmware.api [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Task: {'id': task-1263726, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.257188} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.160250] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1034.160452] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1034.160660] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1034.160880] env[62952]: INFO nova.compute.manager [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1034.161150] env[62952]: DEBUG oslo.service.loopingcall [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1034.161350] env[62952]: DEBUG nova.compute.manager [-] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1034.161440] env[62952]: DEBUG nova.network.neutron [-] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1034.180044] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1034.182839] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-792316b1-0768-4d3c-be8d-8240be2111b7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.184943] env[62952]: DEBUG oslo_concurrency.lockutils [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Acquiring lock "b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.185785] env[62952]: DEBUG oslo_concurrency.lockutils [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Lock "b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.194024] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1034.194024] env[62952]: value = "task-1263728" [ 1034.194024] env[62952]: _type = "Task" [ 1034.194024] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.203445] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263728, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.226898] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.304024] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263727, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5029} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.304024] env[62952]: INFO nova.virt.vmwareapi.ds_util [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] d98a88aa-be5f-4254-aaba-c6fa03c9269a/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk. [ 1034.304024] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a6bd47-032a-463e-842f-772bc0b84dfb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.331074] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] d98a88aa-be5f-4254-aaba-c6fa03c9269a/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1034.331691] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dbdcdb38-72bf-422f-8dee-996429c8f8bc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.352019] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1034.352019] env[62952]: value = "task-1263729" [ 1034.352019] env[62952]: _type = "Task" [ 1034.352019] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.361260] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263729, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.364956] env[62952]: DEBUG oslo_concurrency.lockutils [None req-64e40759-51e3-454f-9a4a-4d71b5a95d4c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "063d1d92-0526-48e0-b33c-f51a79e7c821" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.390040] env[62952]: DEBUG nova.compute.utils [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1034.390040] env[62952]: DEBUG nova.compute.manager [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1034.390404] env[62952]: DEBUG nova.network.neutron [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1034.485042] env[62952]: DEBUG nova.policy [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3ec1aecc0cf64aaea875a4280b114874', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fbb5ccb4d615429facd6932fb55b04e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1034.511287] env[62952]: DEBUG oslo_concurrency.lockutils [req-dc5d31ae-55d5-43d1-87b9-07008de73fcc req-a2bc179d-e834-432c-a26c-6f18a8df8125 service nova] Releasing lock "refresh_cache-063d1d92-0526-48e0-b33c-f51a79e7c821" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.640291] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dbb5c0b-1618-4565-a6c9-18ddb7903597 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.651869] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26549e4b-c0dc-4633-8892-45889cfd26f3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.681203] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a26cce-37cc-47df-b6ee-abb007603b7a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.688664] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9a36200-9613-465c-b287-e1c8cf6dde21 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.692851] env[62952]: DEBUG nova.compute.manager [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1034.706590] env[62952]: DEBUG nova.compute.provider_tree [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.714906] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263728, 'name': PowerOffVM_Task, 'duration_secs': 0.250215} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.714906] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1034.714906] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1034.714906] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75dccd45-1d49-4f48-9d83-9b6b9c439427 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.721529] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1034.721763] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-218a916a-d035-4c64-b9f5-96f19a6d49dd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.796210] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1034.796210] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1034.796423] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Deleting the datastore file [datastore2] 05223053-90c3-433c-99b9-e252ef9a1935 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1034.796566] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c3a940be-1147-4370-8d3d-c634fdbcafe9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.803660] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1034.803660] env[62952]: value = "task-1263731" [ 1034.803660] env[62952]: _type = "Task" [ 1034.803660] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.811291] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263731, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.860134] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263729, 'name': ReconfigVM_Task, 'duration_secs': 0.437859} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.860467] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Reconfigured VM instance instance-0000005a to attach disk [datastore2] d98a88aa-be5f-4254-aaba-c6fa03c9269a/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1034.861389] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f49940-b3fb-4d88-a131-95350e69c22e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.888832] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77ff3533-7c40-4d0f-94ea-df20c2b80b08 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.899197] env[62952]: DEBUG nova.compute.manager [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1034.905043] env[62952]: DEBUG nova.network.neutron [-] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.907464] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1034.907464] env[62952]: value = "task-1263732" [ 1034.907464] env[62952]: _type = "Task" [ 1034.907464] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.915822] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263732, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.013296] env[62952]: DEBUG nova.network.neutron [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Successfully created port: cff4c81c-513a-42a4-b6f5-94b4367fa0ea {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1035.210399] env[62952]: DEBUG nova.scheduler.client.report [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1035.223146] env[62952]: DEBUG oslo_concurrency.lockutils [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.313977] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263731, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.328450] env[62952]: DEBUG nova.compute.manager [req-65d0ed91-44a1-4189-a63d-028076abdb49 req-3512c29f-b095-42bc-a06c-c7c36fd6797c service nova] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Received event network-vif-deleted-272ccc8b-6da5-4832-9304-bcf2814b9283 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1035.407121] env[62952]: INFO nova.compute.manager [-] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Took 1.25 seconds to deallocate network for instance. [ 1035.421296] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263732, 'name': ReconfigVM_Task, 'duration_secs': 0.16394} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.421689] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1035.422110] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e925725a-d3fe-49ad-b573-425b1bd67250 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.428878] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1035.428878] env[62952]: value = "task-1263733" [ 1035.428878] env[62952]: _type = "Task" [ 1035.428878] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.441238] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263733, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.715601] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.829s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.718145] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.351s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.718412] env[62952]: DEBUG nova.objects.instance [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lazy-loading 'resources' on Instance uuid 6f0b04f5-811f-4c53-808b-6d9d22100a86 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.737976] env[62952]: INFO nova.scheduler.client.report [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Deleted allocations for instance 511b995d-c531-41f5-9d1b-89ec8c1bfc18 [ 1035.815828] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263731, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.569268} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.816122] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1035.816321] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1035.816676] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1035.909336] env[62952]: DEBUG nova.compute.manager [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1035.918138] env[62952]: DEBUG oslo_concurrency.lockutils [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.942912] env[62952]: DEBUG oslo_vmware.api [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263733, 'name': PowerOnVM_Task, 'duration_secs': 0.423814} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.945230] env[62952]: DEBUG nova.virt.hardware [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1035.945479] env[62952]: DEBUG nova.virt.hardware [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1035.945663] env[62952]: DEBUG nova.virt.hardware [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1035.945868] env[62952]: DEBUG nova.virt.hardware [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1035.946029] env[62952]: DEBUG nova.virt.hardware [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1035.946183] env[62952]: DEBUG nova.virt.hardware [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1035.946398] env[62952]: DEBUG nova.virt.hardware [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1035.946559] env[62952]: DEBUG nova.virt.hardware [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1035.947372] env[62952]: DEBUG nova.virt.hardware [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1035.947372] env[62952]: DEBUG nova.virt.hardware [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1035.947372] env[62952]: DEBUG nova.virt.hardware [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1035.947372] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1035.949547] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b48259-152c-4ad2-be04-704f90970a2c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.952909] env[62952]: DEBUG nova.compute.manager [None req-c3e42627-8467-4e49-ac77-52fd106a86f7 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1035.953691] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e230e97-9d8b-4581-b95e-f8b440651eb1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.962898] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3a1b49-3c7f-4667-a73f-537b8708282b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.246242] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38b2946f-32ae-478e-85de-c63cd10db988 tempest-ServerMetadataNegativeTestJSON-449513043 tempest-ServerMetadataNegativeTestJSON-449513043-project-member] Lock "511b995d-c531-41f5-9d1b-89ec8c1bfc18" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.467s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.404337] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ddc9566-c7ad-4add-a813-33df40431645 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.412288] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e98461d1-4d8e-4652-ae49-fecf851313f7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.443726] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d67324-38f0-4858-b174-c27134b8deb6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.450759] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f9918b-a3ca-4ee0-8675-b69a29a763e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.464230] env[62952]: DEBUG nova.compute.provider_tree [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1036.832059] env[62952]: INFO nova.compute.manager [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Unrescuing [ 1036.832303] env[62952]: DEBUG oslo_concurrency.lockutils [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "refresh_cache-d98a88aa-be5f-4254-aaba-c6fa03c9269a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.832499] env[62952]: DEBUG oslo_concurrency.lockutils [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquired lock "refresh_cache-d98a88aa-be5f-4254-aaba-c6fa03c9269a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.832680] env[62952]: DEBUG nova.network.neutron [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1036.855850] env[62952]: DEBUG nova.virt.hardware [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1036.856132] env[62952]: DEBUG nova.virt.hardware [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1036.856300] env[62952]: DEBUG nova.virt.hardware [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1036.856484] env[62952]: DEBUG nova.virt.hardware [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1036.856636] env[62952]: DEBUG nova.virt.hardware [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1036.856832] env[62952]: DEBUG nova.virt.hardware [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1036.857426] env[62952]: DEBUG nova.virt.hardware [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1036.857776] env[62952]: DEBUG nova.virt.hardware [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1036.857997] env[62952]: DEBUG nova.virt.hardware [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1036.858188] env[62952]: DEBUG nova.virt.hardware [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1036.858429] env[62952]: DEBUG nova.virt.hardware [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1036.859326] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6de23f0-3aad-45aa-8185-5b048df2ffe6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.868782] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c28ec282-0a9a-4466-921a-97d45416cb7c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.883322] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:43:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e1c9bb98-73a9-48eb-856e-a541afe9b07b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fe560683-a241-423f-9f16-c5f59668d967', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1036.890924] env[62952]: DEBUG oslo.service.loopingcall [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.891185] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1036.891424] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cd2ed930-a0e9-4fa3-8f95-83f4a753ddc7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.909649] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1036.909649] env[62952]: value = "task-1263734" [ 1036.909649] env[62952]: _type = "Task" [ 1036.909649] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.918201] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263734, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.967786] env[62952]: DEBUG nova.scheduler.client.report [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1037.060317] env[62952]: DEBUG nova.compute.manager [req-2efcdbee-9e61-4390-a264-9ce3ee58ab96 req-7d2fb27d-a1a7-49a7-af8c-2de6d10e82ab service nova] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Received event network-vif-plugged-cff4c81c-513a-42a4-b6f5-94b4367fa0ea {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1037.060631] env[62952]: DEBUG oslo_concurrency.lockutils [req-2efcdbee-9e61-4390-a264-9ce3ee58ab96 req-7d2fb27d-a1a7-49a7-af8c-2de6d10e82ab service nova] Acquiring lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.061752] env[62952]: DEBUG oslo_concurrency.lockutils [req-2efcdbee-9e61-4390-a264-9ce3ee58ab96 req-7d2fb27d-a1a7-49a7-af8c-2de6d10e82ab service nova] Lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.061882] env[62952]: DEBUG oslo_concurrency.lockutils [req-2efcdbee-9e61-4390-a264-9ce3ee58ab96 req-7d2fb27d-a1a7-49a7-af8c-2de6d10e82ab service nova] Lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.062236] env[62952]: DEBUG nova.compute.manager [req-2efcdbee-9e61-4390-a264-9ce3ee58ab96 req-7d2fb27d-a1a7-49a7-af8c-2de6d10e82ab service nova] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] No waiting events found dispatching network-vif-plugged-cff4c81c-513a-42a4-b6f5-94b4367fa0ea {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1037.062335] env[62952]: WARNING nova.compute.manager [req-2efcdbee-9e61-4390-a264-9ce3ee58ab96 req-7d2fb27d-a1a7-49a7-af8c-2de6d10e82ab service nova] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Received unexpected event network-vif-plugged-cff4c81c-513a-42a4-b6f5-94b4367fa0ea for instance with vm_state building and task_state spawning. [ 1037.321178] env[62952]: DEBUG nova.network.neutron [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Successfully updated port: cff4c81c-513a-42a4-b6f5-94b4367fa0ea {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1037.329142] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1037.329450] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1037.419800] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263734, 'name': CreateVM_Task, 'duration_secs': 0.337685} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.420128] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1037.420653] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.420825] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.421176] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1037.421426] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44ed31ad-71c7-45d9-9068-e765c00ec877 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.425943] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1037.425943] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5220d617-283b-86cf-7ce0-37def311e859" [ 1037.425943] env[62952]: _type = "Task" [ 1037.425943] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.435341] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5220d617-283b-86cf-7ce0-37def311e859, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.475198] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.757s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.477418] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.809s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.477611] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.479290] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.925s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.479553] env[62952]: DEBUG nova.objects.instance [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lazy-loading 'resources' on Instance uuid 063d1d92-0526-48e0-b33c-f51a79e7c821 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.505628] env[62952]: INFO nova.scheduler.client.report [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleted allocations for instance 6f0b04f5-811f-4c53-808b-6d9d22100a86 [ 1037.514060] env[62952]: INFO nova.scheduler.client.report [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleted allocations for instance 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8 [ 1037.757216] env[62952]: DEBUG nova.network.neutron [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Updating instance_info_cache with network_info: [{"id": "3dff9ce5-4a76-49a1-ac02-2702ccde508f", "address": "fa:16:3e:95:0b:fe", "network": {"id": "d84b900e-d793-4187-aaeb-bff9b040e607", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1370737473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a6fddef5a2114acb9749b1bae171cbca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dff9ce5-4a", "ovs_interfaceid": "3dff9ce5-4a76-49a1-ac02-2702ccde508f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.824159] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "refresh_cache-e9e690b8-2fae-401d-aaad-0acef6e88ce3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.824319] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "refresh_cache-e9e690b8-2fae-401d-aaad-0acef6e88ce3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.824474] env[62952]: DEBUG nova.network.neutron [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1037.837108] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1037.837108] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1037.936779] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5220d617-283b-86cf-7ce0-37def311e859, 'name': SearchDatastore_Task, 'duration_secs': 0.019984} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.937110] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.937377] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1037.937711] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.937812] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.937998] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1037.938286] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c7c965e-7310-4ce6-ba0a-bc81af118524 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.947407] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1037.947678] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1037.948480] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7e050da-3196-48fc-894e-26dd63e6d8cb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.960991] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1037.960991] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5223e235-0d32-33b1-68f0-c57f64efeed3" [ 1037.960991] env[62952]: _type = "Task" [ 1037.960991] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.971166] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5223e235-0d32-33b1-68f0-c57f64efeed3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.984770] env[62952]: DEBUG nova.objects.instance [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lazy-loading 'numa_topology' on Instance uuid 063d1d92-0526-48e0-b33c-f51a79e7c821 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.020161] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c3736752-0c64-4b47-8af6-d8f3969336fe tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "6f0b04f5-811f-4c53-808b-6d9d22100a86" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.596s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.023746] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91793f6b-2e8f-4d4e-9591-22c6a6b0adfb tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.898s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.259966] env[62952]: DEBUG oslo_concurrency.lockutils [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Releasing lock "refresh_cache-d98a88aa-be5f-4254-aaba-c6fa03c9269a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.260655] env[62952]: DEBUG nova.objects.instance [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lazy-loading 'flavor' on Instance uuid d98a88aa-be5f-4254-aaba-c6fa03c9269a {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.366407] env[62952]: DEBUG nova.network.neutron [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1038.475434] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5223e235-0d32-33b1-68f0-c57f64efeed3, 'name': SearchDatastore_Task, 'duration_secs': 0.014682} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.475434] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b87b4cae-5ac2-4ded-8bd6-ef0313897aa8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.481720] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1038.481720] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523cb596-85f1-11c1-cad9-f2d6f58fa07b" [ 1038.481720] env[62952]: _type = "Task" [ 1038.481720] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.487650] env[62952]: DEBUG nova.objects.base [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Object Instance<063d1d92-0526-48e0-b33c-f51a79e7c821> lazy-loaded attributes: resources,numa_topology {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1038.495707] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523cb596-85f1-11c1-cad9-f2d6f58fa07b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.580568] env[62952]: DEBUG nova.network.neutron [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Updating instance_info_cache with network_info: [{"id": "cff4c81c-513a-42a4-b6f5-94b4367fa0ea", "address": "fa:16:3e:3a:e4:22", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcff4c81c-51", "ovs_interfaceid": "cff4c81c-513a-42a4-b6f5-94b4367fa0ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.687086] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d5753c-5cce-4329-bb3e-fdeef9d0958a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.695385] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4adef620-3720-4160-9f96-e077ba5c331a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.727626] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44ab2c39-67e0-4619-9779-ae019c3f1538 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.735553] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac84224-ce68-4460-97fa-0a5930d40a6e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.750783] env[62952]: DEBUG nova.compute.provider_tree [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1038.767764] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee4e045-26c5-46ac-b2d7-ec3eb0503e2b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.790124] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1038.790366] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-738d5b15-e190-4c33-885e-eecf2d6ba1ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.796723] env[62952]: DEBUG oslo_vmware.api [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1038.796723] env[62952]: value = "task-1263735" [ 1038.796723] env[62952]: _type = "Task" [ 1038.796723] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.804875] env[62952]: DEBUG oslo_vmware.api [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263735, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.994292] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523cb596-85f1-11c1-cad9-f2d6f58fa07b, 'name': SearchDatastore_Task, 'duration_secs': 0.009353} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.994790] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.996310] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 05223053-90c3-433c-99b9-e252ef9a1935/05223053-90c3-433c-99b9-e252ef9a1935.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1038.996310] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d6145c94-6819-4a62-9789-9b2aac3af029 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.004151] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1039.004151] env[62952]: value = "task-1263736" [ 1039.004151] env[62952]: _type = "Task" [ 1039.004151] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.012979] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263736, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.085025] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "refresh_cache-e9e690b8-2fae-401d-aaad-0acef6e88ce3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.085025] env[62952]: DEBUG nova.compute.manager [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Instance network_info: |[{"id": "cff4c81c-513a-42a4-b6f5-94b4367fa0ea", "address": "fa:16:3e:3a:e4:22", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcff4c81c-51", "ovs_interfaceid": "cff4c81c-513a-42a4-b6f5-94b4367fa0ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1039.085025] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:e4:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5ad8894f-e240-4013-8272-4e79daea0751', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cff4c81c-513a-42a4-b6f5-94b4367fa0ea', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1039.092793] env[62952]: DEBUG oslo.service.loopingcall [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.093053] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1039.093288] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4090f046-ed65-4dc0-95f2-c72f5bc27eae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.111844] env[62952]: DEBUG nova.compute.manager [req-d25c1283-2a6f-41b9-9368-3318ccde0829 req-9a2b48c4-1108-4682-bd7b-3ae58ba74ef6 service nova] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Received event network-changed-cff4c81c-513a-42a4-b6f5-94b4367fa0ea {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1039.112054] env[62952]: DEBUG nova.compute.manager [req-d25c1283-2a6f-41b9-9368-3318ccde0829 req-9a2b48c4-1108-4682-bd7b-3ae58ba74ef6 service nova] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Refreshing instance network info cache due to event network-changed-cff4c81c-513a-42a4-b6f5-94b4367fa0ea. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1039.112271] env[62952]: DEBUG oslo_concurrency.lockutils [req-d25c1283-2a6f-41b9-9368-3318ccde0829 req-9a2b48c4-1108-4682-bd7b-3ae58ba74ef6 service nova] Acquiring lock "refresh_cache-e9e690b8-2fae-401d-aaad-0acef6e88ce3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.112416] env[62952]: DEBUG oslo_concurrency.lockutils [req-d25c1283-2a6f-41b9-9368-3318ccde0829 req-9a2b48c4-1108-4682-bd7b-3ae58ba74ef6 service nova] Acquired lock "refresh_cache-e9e690b8-2fae-401d-aaad-0acef6e88ce3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.112626] env[62952]: DEBUG nova.network.neutron [req-d25c1283-2a6f-41b9-9368-3318ccde0829 req-9a2b48c4-1108-4682-bd7b-3ae58ba74ef6 service nova] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Refreshing network info cache for port cff4c81c-513a-42a4-b6f5-94b4367fa0ea {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1039.120246] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1039.120246] env[62952]: value = "task-1263737" [ 1039.120246] env[62952]: _type = "Task" [ 1039.120246] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.129353] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263737, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.254044] env[62952]: DEBUG nova.scheduler.client.report [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.307812] env[62952]: DEBUG oslo_vmware.api [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263735, 'name': PowerOffVM_Task, 'duration_secs': 0.224296} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.308304] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1039.314082] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Reconfiguring VM instance instance-0000005a to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1039.314870] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1aec8074-3e76-441b-8f9b-b5cd7137dc48 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.334393] env[62952]: DEBUG oslo_vmware.api [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1039.334393] env[62952]: value = "task-1263738" [ 1039.334393] env[62952]: _type = "Task" [ 1039.334393] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.343456] env[62952]: DEBUG oslo_vmware.api [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263738, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.396579] env[62952]: DEBUG oslo_concurrency.lockutils [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "1a38bdcd-5537-46f5-94f6-a8e358e13121" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.396808] env[62952]: DEBUG oslo_concurrency.lockutils [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "1a38bdcd-5537-46f5-94f6-a8e358e13121" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.514431] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263736, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.459889} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.514788] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 05223053-90c3-433c-99b9-e252ef9a1935/05223053-90c3-433c-99b9-e252ef9a1935.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1039.514905] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1039.515172] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7636a33d-af63-44ec-b835-dbceb1f17ff7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.520734] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1039.520734] env[62952]: value = "task-1263739" [ 1039.520734] env[62952]: _type = "Task" [ 1039.520734] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.529732] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263739, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.629521] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263737, 'name': CreateVM_Task, 'duration_secs': 0.359235} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.629738] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1039.630424] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.630594] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.630919] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1039.631191] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37473cb0-7f63-4e80-a8ce-9369d8385aec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.635529] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1039.635529] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52be75d8-5fc4-3a0a-a854-93d30580baaa" [ 1039.635529] env[62952]: _type = "Task" [ 1039.635529] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.645325] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52be75d8-5fc4-3a0a-a854-93d30580baaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.762836] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.283s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.766559] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.539s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.767377] env[62952]: INFO nova.compute.claims [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1039.821546] env[62952]: DEBUG nova.network.neutron [req-d25c1283-2a6f-41b9-9368-3318ccde0829 req-9a2b48c4-1108-4682-bd7b-3ae58ba74ef6 service nova] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Updated VIF entry in instance network info cache for port cff4c81c-513a-42a4-b6f5-94b4367fa0ea. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1039.821912] env[62952]: DEBUG nova.network.neutron [req-d25c1283-2a6f-41b9-9368-3318ccde0829 req-9a2b48c4-1108-4682-bd7b-3ae58ba74ef6 service nova] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Updating instance_info_cache with network_info: [{"id": "cff4c81c-513a-42a4-b6f5-94b4367fa0ea", "address": "fa:16:3e:3a:e4:22", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcff4c81c-51", "ovs_interfaceid": "cff4c81c-513a-42a4-b6f5-94b4367fa0ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.844666] env[62952]: DEBUG oslo_vmware.api [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263738, 'name': ReconfigVM_Task, 'duration_secs': 0.259166} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.844975] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Reconfigured VM instance instance-0000005a to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1039.845261] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1039.845557] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7bfe5333-3e79-4db7-8451-3dae71c44526 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.855490] env[62952]: DEBUG oslo_vmware.api [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1039.855490] env[62952]: value = "task-1263740" [ 1039.855490] env[62952]: _type = "Task" [ 1039.855490] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.865870] env[62952]: DEBUG oslo_vmware.api [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263740, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.892855] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.893020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquired lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.893180] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Forcefully refreshing network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1039.899485] env[62952]: DEBUG nova.compute.manager [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1040.031274] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263739, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059956} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.031555] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1040.032534] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-957a0944-9d64-4085-9565-2ad764da8feb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.055404] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 05223053-90c3-433c-99b9-e252ef9a1935/05223053-90c3-433c-99b9-e252ef9a1935.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1040.055858] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06aa8ae7-02dd-48c7-9b96-5278fd9135f7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.076769] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1040.076769] env[62952]: value = "task-1263741" [ 1040.076769] env[62952]: _type = "Task" [ 1040.076769] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.084859] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263741, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.144916] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52be75d8-5fc4-3a0a-a854-93d30580baaa, 'name': SearchDatastore_Task, 'duration_secs': 0.009404} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.145262] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.145558] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1040.145922] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.146190] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.146481] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1040.146883] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-94d3fb03-c797-4761-b45a-3beaeb23fcb8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.155312] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1040.155491] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1040.156252] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e5bb60a-57af-4177-8df4-19e592dcd288 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.161346] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1040.161346] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]522dcd6e-6de0-e95e-5f66-342dda320b50" [ 1040.161346] env[62952]: _type = "Task" [ 1040.161346] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.168644] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]522dcd6e-6de0-e95e-5f66-342dda320b50, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.272430] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5ca37ae1-3940-4884-85fe-2e46310ecffb tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "063d1d92-0526-48e0-b33c-f51a79e7c821" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 28.723s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.273261] env[62952]: DEBUG oslo_concurrency.lockutils [None req-64e40759-51e3-454f-9a4a-4d71b5a95d4c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "063d1d92-0526-48e0-b33c-f51a79e7c821" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 5.908s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.273486] env[62952]: DEBUG oslo_concurrency.lockutils [None req-64e40759-51e3-454f-9a4a-4d71b5a95d4c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "063d1d92-0526-48e0-b33c-f51a79e7c821-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.273699] env[62952]: DEBUG oslo_concurrency.lockutils [None req-64e40759-51e3-454f-9a4a-4d71b5a95d4c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "063d1d92-0526-48e0-b33c-f51a79e7c821-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.273869] env[62952]: DEBUG oslo_concurrency.lockutils [None req-64e40759-51e3-454f-9a4a-4d71b5a95d4c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "063d1d92-0526-48e0-b33c-f51a79e7c821-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.276818] env[62952]: INFO nova.compute.manager [None req-64e40759-51e3-454f-9a4a-4d71b5a95d4c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Terminating instance [ 1040.278563] env[62952]: DEBUG nova.compute.manager [None req-64e40759-51e3-454f-9a4a-4d71b5a95d4c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1040.278761] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-64e40759-51e3-454f-9a4a-4d71b5a95d4c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1040.279026] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a633489a-6bea-48b5-ae35-d105832b3c38 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.288679] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fab3f41-bf36-45c2-b9e4-b224df543ed1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.322014] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-64e40759-51e3-454f-9a4a-4d71b5a95d4c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 063d1d92-0526-48e0-b33c-f51a79e7c821 could not be found. [ 1040.322267] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-64e40759-51e3-454f-9a4a-4d71b5a95d4c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1040.322463] env[62952]: INFO nova.compute.manager [None req-64e40759-51e3-454f-9a4a-4d71b5a95d4c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Took 0.04 seconds to destroy the instance on the hypervisor. [ 1040.322719] env[62952]: DEBUG oslo.service.loopingcall [None req-64e40759-51e3-454f-9a4a-4d71b5a95d4c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1040.322976] env[62952]: DEBUG nova.compute.manager [-] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1040.323081] env[62952]: DEBUG nova.network.neutron [-] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1040.324993] env[62952]: DEBUG oslo_concurrency.lockutils [req-d25c1283-2a6f-41b9-9368-3318ccde0829 req-9a2b48c4-1108-4682-bd7b-3ae58ba74ef6 service nova] Releasing lock "refresh_cache-e9e690b8-2fae-401d-aaad-0acef6e88ce3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.367784] env[62952]: DEBUG oslo_vmware.api [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263740, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.423157] env[62952]: DEBUG oslo_concurrency.lockutils [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.587954] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263741, 'name': ReconfigVM_Task, 'duration_secs': 0.265832} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.588359] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 05223053-90c3-433c-99b9-e252ef9a1935/05223053-90c3-433c-99b9-e252ef9a1935.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1040.589110] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7770478f-3071-4ea3-864f-41c5a758d304 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.595782] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1040.595782] env[62952]: value = "task-1263742" [ 1040.595782] env[62952]: _type = "Task" [ 1040.595782] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.606320] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263742, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.674599] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]522dcd6e-6de0-e95e-5f66-342dda320b50, 'name': SearchDatastore_Task, 'duration_secs': 0.008371} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.675126] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6158e3e-4d01-483d-9924-26063105cd80 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.680141] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1040.680141] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520c6ce0-cd6b-6cf4-7ad9-73ea261cdce0" [ 1040.680141] env[62952]: _type = "Task" [ 1040.680141] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.687937] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520c6ce0-cd6b-6cf4-7ad9-73ea261cdce0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.866754] env[62952]: DEBUG oslo_vmware.api [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263740, 'name': PowerOnVM_Task, 'duration_secs': 0.548501} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.869453] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1040.869753] env[62952]: DEBUG nova.compute.manager [None req-05e41dd3-0659-4480-8cee-fc00767b3ff4 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1040.870871] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7fa2922-0b5f-4b81-a92a-c78b1c04557b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.930877] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e62d0d-f6d9-4453-89e7-30b6ba9643ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.938833] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e02a23-34c4-4457-8c8c-ff17daebd4db {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.970310] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-858ac396-7c13-4582-bf4e-5eac3d43636a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.979251] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f0acd3-f528-44bd-8437-9bd5b675b7ea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.992489] env[62952]: DEBUG nova.compute.provider_tree [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.106134] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263742, 'name': Rename_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.140903] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance_info_cache with network_info: [{"id": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "address": "fa:16:3e:eb:c2:8d", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e2e9d1e-dc", "ovs_interfaceid": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.149470] env[62952]: DEBUG nova.network.neutron [-] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.196733] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520c6ce0-cd6b-6cf4-7ad9-73ea261cdce0, 'name': SearchDatastore_Task, 'duration_secs': 0.011008} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.196733] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.196733] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] e9e690b8-2fae-401d-aaad-0acef6e88ce3/e9e690b8-2fae-401d-aaad-0acef6e88ce3.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1041.196733] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-05ed3148-825b-4073-aabf-0733007956e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.201877] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1041.201877] env[62952]: value = "task-1263743" [ 1041.201877] env[62952]: _type = "Task" [ 1041.201877] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.210455] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263743, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.496090] env[62952]: DEBUG nova.scheduler.client.report [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.609330] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263742, 'name': Rename_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.644190] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Releasing lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.644351] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updated the network info_cache for instance {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1041.644579] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.644819] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.645017] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.645238] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.645404] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.645555] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.645680] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1041.645822] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.651738] env[62952]: INFO nova.compute.manager [-] [instance: 063d1d92-0526-48e0-b33c-f51a79e7c821] Took 1.33 seconds to deallocate network for instance. [ 1041.711196] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263743, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446649} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.711503] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] e9e690b8-2fae-401d-aaad-0acef6e88ce3/e9e690b8-2fae-401d-aaad-0acef6e88ce3.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1041.711733] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1041.711990] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7af76bc1-d330-4e08-aaec-502801ff507c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.718417] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1041.718417] env[62952]: value = "task-1263744" [ 1041.718417] env[62952]: _type = "Task" [ 1041.718417] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.725645] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263744, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.802171] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "d98a88aa-be5f-4254-aaba-c6fa03c9269a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.802455] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "d98a88aa-be5f-4254-aaba-c6fa03c9269a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.802675] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "d98a88aa-be5f-4254-aaba-c6fa03c9269a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.802853] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "d98a88aa-be5f-4254-aaba-c6fa03c9269a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.803034] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "d98a88aa-be5f-4254-aaba-c6fa03c9269a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.805315] env[62952]: INFO nova.compute.manager [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Terminating instance [ 1041.807009] env[62952]: DEBUG nova.compute.manager [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1041.807229] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1041.808054] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c9c25d-8ca9-4b42-9a4e-2bfe4b859f6d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.815322] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1041.815805] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-499a5516-5e5a-4c39-8d39-dd22193746f5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.821812] env[62952]: DEBUG oslo_vmware.api [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1041.821812] env[62952]: value = "task-1263745" [ 1041.821812] env[62952]: _type = "Task" [ 1041.821812] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.829212] env[62952]: DEBUG oslo_vmware.api [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263745, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.002381] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.236s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.002937] env[62952]: DEBUG nova.compute.manager [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1042.005724] env[62952]: DEBUG oslo_concurrency.lockutils [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.783s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.007221] env[62952]: INFO nova.compute.claims [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1042.106557] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263742, 'name': Rename_Task, 'duration_secs': 1.137548} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.106827] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1042.107366] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5fd5dbce-93a8-4668-a5c4-e002af03c7e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.113411] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1042.113411] env[62952]: value = "task-1263746" [ 1042.113411] env[62952]: _type = "Task" [ 1042.113411] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.120273] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263746, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.149323] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.228449] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263744, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067346} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.228743] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1042.229629] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-216d2f33-1f7a-4edc-93ab-382ebca69e14 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.251104] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] e9e690b8-2fae-401d-aaad-0acef6e88ce3/e9e690b8-2fae-401d-aaad-0acef6e88ce3.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1042.251618] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5979477b-7168-4ef6-8cad-96caddaa0905 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.270396] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1042.270396] env[62952]: value = "task-1263747" [ 1042.270396] env[62952]: _type = "Task" [ 1042.270396] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.277769] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263747, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.333232] env[62952]: DEBUG oslo_vmware.api [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263745, 'name': PowerOffVM_Task, 'duration_secs': 0.210314} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.333516] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1042.333687] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1042.333933] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8b417622-198b-471a-9a87-05133edccf27 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.393390] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1042.393679] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1042.393910] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Deleting the datastore file [datastore2] d98a88aa-be5f-4254-aaba-c6fa03c9269a {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1042.394357] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-87f4a4ea-e5a6-4885-8cca-f3276ae2b371 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.401389] env[62952]: DEBUG oslo_vmware.api [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1042.401389] env[62952]: value = "task-1263749" [ 1042.401389] env[62952]: _type = "Task" [ 1042.401389] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.410184] env[62952]: DEBUG oslo_vmware.api [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263749, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.511782] env[62952]: DEBUG nova.compute.utils [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1042.515663] env[62952]: DEBUG nova.compute.manager [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1042.515838] env[62952]: DEBUG nova.network.neutron [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1042.565269] env[62952]: DEBUG nova.policy [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '42a603de55734dff95126c1c9bcc8699', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd93803f89ca64e66b8271e7b0bb48983', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1042.622625] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263746, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.677369] env[62952]: DEBUG oslo_concurrency.lockutils [None req-64e40759-51e3-454f-9a4a-4d71b5a95d4c tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "063d1d92-0526-48e0-b33c-f51a79e7c821" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.403s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.780187] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263747, 'name': ReconfigVM_Task, 'duration_secs': 0.423553} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.780567] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Reconfigured VM instance instance-0000005d to attach disk [datastore1] e9e690b8-2fae-401d-aaad-0acef6e88ce3/e9e690b8-2fae-401d-aaad-0acef6e88ce3.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1042.781198] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-420efb97-b10e-44c1-a0da-316c43429c87 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.788127] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1042.788127] env[62952]: value = "task-1263750" [ 1042.788127] env[62952]: _type = "Task" [ 1042.788127] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.795740] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263750, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.910581] env[62952]: DEBUG oslo_vmware.api [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263749, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159978} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.910859] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1042.911061] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1042.911252] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1042.911436] env[62952]: INFO nova.compute.manager [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1042.911779] env[62952]: DEBUG oslo.service.loopingcall [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.911861] env[62952]: DEBUG nova.compute.manager [-] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1042.912030] env[62952]: DEBUG nova.network.neutron [-] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1042.973635] env[62952]: DEBUG nova.network.neutron [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Successfully created port: ddd93d49-eb57-4af9-a9bb-35d8269b939d {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1043.017548] env[62952]: DEBUG nova.compute.manager [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1043.126029] env[62952]: DEBUG oslo_vmware.api [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263746, 'name': PowerOnVM_Task, 'duration_secs': 0.514335} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.128602] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1043.128813] env[62952]: DEBUG nova.compute.manager [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1043.129754] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb489f32-1daf-4c30-9730-18377d0ef29f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.201029] env[62952]: DEBUG nova.compute.manager [req-9f50900a-87b7-4954-88f3-7685bdc058a4 req-b271a8e8-7716-47c7-ad4f-c314b933a464 service nova] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Received event network-vif-deleted-3dff9ce5-4a76-49a1-ac02-2702ccde508f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1043.201029] env[62952]: INFO nova.compute.manager [req-9f50900a-87b7-4954-88f3-7685bdc058a4 req-b271a8e8-7716-47c7-ad4f-c314b933a464 service nova] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Neutron deleted interface 3dff9ce5-4a76-49a1-ac02-2702ccde508f; detaching it from the instance and deleting it from the info cache [ 1043.201206] env[62952]: DEBUG nova.network.neutron [req-9f50900a-87b7-4954-88f3-7685bdc058a4 req-b271a8e8-7716-47c7-ad4f-c314b933a464 service nova] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.230734] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ae08029-b14c-4510-9ab6-83b9f6733979 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.239642] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a86e41fe-18bc-4aa1-b7c1-e50b4c05be29 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.271374] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85016d2b-0916-403b-9625-a9e78d4236c9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.279850] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b908a56-1395-4956-84af-00e0f999524c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.293124] env[62952]: DEBUG nova.compute.provider_tree [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.302191] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263750, 'name': Rename_Task, 'duration_secs': 0.325988} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.303124] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1043.303370] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cc762c7f-55ec-4c80-902d-abd4606b5a99 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.309457] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1043.309457] env[62952]: value = "task-1263751" [ 1043.309457] env[62952]: _type = "Task" [ 1043.309457] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.317341] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263751, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.646849] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.681555] env[62952]: DEBUG nova.network.neutron [-] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.703814] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-243ebe16-5c33-4b1b-a9cf-6a85886183fb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.713950] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e78d7ec-4985-4178-8647-cde4b200fd4f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.743604] env[62952]: DEBUG nova.compute.manager [req-9f50900a-87b7-4954-88f3-7685bdc058a4 req-b271a8e8-7716-47c7-ad4f-c314b933a464 service nova] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Detach interface failed, port_id=3dff9ce5-4a76-49a1-ac02-2702ccde508f, reason: Instance d98a88aa-be5f-4254-aaba-c6fa03c9269a could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1043.796316] env[62952]: DEBUG nova.scheduler.client.report [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1043.822304] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263751, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.034023] env[62952]: DEBUG nova.compute.manager [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1044.058369] env[62952]: DEBUG nova.virt.hardware [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1044.058587] env[62952]: DEBUG nova.virt.hardware [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1044.058745] env[62952]: DEBUG nova.virt.hardware [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1044.058986] env[62952]: DEBUG nova.virt.hardware [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1044.059221] env[62952]: DEBUG nova.virt.hardware [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1044.059397] env[62952]: DEBUG nova.virt.hardware [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1044.059610] env[62952]: DEBUG nova.virt.hardware [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1044.060104] env[62952]: DEBUG nova.virt.hardware [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1044.060362] env[62952]: DEBUG nova.virt.hardware [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1044.060545] env[62952]: DEBUG nova.virt.hardware [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1044.060721] env[62952]: DEBUG nova.virt.hardware [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1044.061592] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c193e6-ebf4-472e-a89e-4e6d7a283767 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.069830] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dbcb4a5-ce5e-4a9f-81b9-e87a2989f32f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.184626] env[62952]: INFO nova.compute.manager [-] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Took 1.27 seconds to deallocate network for instance. [ 1044.200829] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "af261b32-457a-44a1-bab4-b62654b0be1f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.201730] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "af261b32-457a-44a1-bab4-b62654b0be1f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.302058] env[62952]: DEBUG oslo_concurrency.lockutils [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.296s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.302403] env[62952]: DEBUG nova.compute.manager [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1044.305251] env[62952]: DEBUG oslo_concurrency.lockutils [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.387s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.305735] env[62952]: DEBUG nova.objects.instance [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Lazy-loading 'resources' on Instance uuid 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.320498] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263751, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.438120] env[62952]: DEBUG nova.compute.manager [req-6b21b498-5229-481a-bfdc-b35c1b68fa9c req-5e751ba7-ddb5-4c43-86e0-1bda576c6a20 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Received event network-vif-plugged-ddd93d49-eb57-4af9-a9bb-35d8269b939d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1044.438328] env[62952]: DEBUG oslo_concurrency.lockutils [req-6b21b498-5229-481a-bfdc-b35c1b68fa9c req-5e751ba7-ddb5-4c43-86e0-1bda576c6a20 service nova] Acquiring lock "42793459-89e6-4941-9b20-cbe52a241c6d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.438557] env[62952]: DEBUG oslo_concurrency.lockutils [req-6b21b498-5229-481a-bfdc-b35c1b68fa9c req-5e751ba7-ddb5-4c43-86e0-1bda576c6a20 service nova] Lock "42793459-89e6-4941-9b20-cbe52a241c6d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.438756] env[62952]: DEBUG oslo_concurrency.lockutils [req-6b21b498-5229-481a-bfdc-b35c1b68fa9c req-5e751ba7-ddb5-4c43-86e0-1bda576c6a20 service nova] Lock "42793459-89e6-4941-9b20-cbe52a241c6d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.438930] env[62952]: DEBUG nova.compute.manager [req-6b21b498-5229-481a-bfdc-b35c1b68fa9c req-5e751ba7-ddb5-4c43-86e0-1bda576c6a20 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] No waiting events found dispatching network-vif-plugged-ddd93d49-eb57-4af9-a9bb-35d8269b939d {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1044.439145] env[62952]: WARNING nova.compute.manager [req-6b21b498-5229-481a-bfdc-b35c1b68fa9c req-5e751ba7-ddb5-4c43-86e0-1bda576c6a20 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Received unexpected event network-vif-plugged-ddd93d49-eb57-4af9-a9bb-35d8269b939d for instance with vm_state building and task_state spawning. [ 1044.595236] env[62952]: DEBUG nova.network.neutron [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Successfully updated port: ddd93d49-eb57-4af9-a9bb-35d8269b939d {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1044.692213] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.703274] env[62952]: DEBUG nova.compute.manager [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1044.812461] env[62952]: DEBUG nova.compute.utils [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1044.814443] env[62952]: DEBUG nova.compute.manager [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1044.815331] env[62952]: DEBUG nova.network.neutron [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1044.825906] env[62952]: DEBUG oslo_vmware.api [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263751, 'name': PowerOnVM_Task, 'duration_secs': 1.226592} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.826688] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1044.826903] env[62952]: INFO nova.compute.manager [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Took 8.92 seconds to spawn the instance on the hypervisor. [ 1044.827173] env[62952]: DEBUG nova.compute.manager [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1044.828052] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd0b2abb-d25d-4d23-98b1-b8c119a1cbf0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.861158] env[62952]: DEBUG nova.policy [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '399de9607a214f2db8e4bbc89d9aa4b2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bff36f8c24b44ea8afc9094153087c5c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1044.908441] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "05223053-90c3-433c-99b9-e252ef9a1935" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.908708] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "05223053-90c3-433c-99b9-e252ef9a1935" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.908915] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "05223053-90c3-433c-99b9-e252ef9a1935-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.909112] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "05223053-90c3-433c-99b9-e252ef9a1935-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.910047] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "05223053-90c3-433c-99b9-e252ef9a1935-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.911680] env[62952]: INFO nova.compute.manager [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Terminating instance [ 1044.913420] env[62952]: DEBUG nova.compute.manager [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1044.913618] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1044.914517] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41ff0b8-1be7-4d7e-8ec3-7ffbbdef0d6d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.924688] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1044.924939] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c175790e-e4d8-4c95-9adc-bebf6cd0b029 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.933741] env[62952]: DEBUG oslo_vmware.api [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1044.933741] env[62952]: value = "task-1263752" [ 1044.933741] env[62952]: _type = "Task" [ 1044.933741] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.944009] env[62952]: DEBUG oslo_vmware.api [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263752, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.997765] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ab6e18-70ec-4370-b6f1-18efb2accae2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.007055] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be52452-c2e5-4eca-b2ea-3aa00dc211f7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.037456] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2649776b-3917-4ef4-8fe0-28f808a80ff2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.045250] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d164d383-b603-4620-9ce6-81d393607fa4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.058300] env[62952]: DEBUG nova.compute.provider_tree [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.101768] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.102130] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.102130] env[62952]: DEBUG nova.network.neutron [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1045.221920] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.318269] env[62952]: DEBUG nova.compute.manager [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1045.352358] env[62952]: INFO nova.compute.manager [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Took 14.67 seconds to build instance. [ 1045.443700] env[62952]: DEBUG oslo_vmware.api [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263752, 'name': PowerOffVM_Task, 'duration_secs': 0.245524} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.444119] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1045.444358] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1045.444677] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d82dd865-5f1a-424a-964b-6d54370fd4ea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.561142] env[62952]: DEBUG nova.scheduler.client.report [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1045.649024] env[62952]: DEBUG nova.network.neutron [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1045.667730] env[62952]: DEBUG nova.network.neutron [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Successfully created port: d5013042-c6e4-4118-aaa6-3947b89a6fb3 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1045.816454] env[62952]: DEBUG nova.network.neutron [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Updating instance_info_cache with network_info: [{"id": "ddd93d49-eb57-4af9-a9bb-35d8269b939d", "address": "fa:16:3e:e3:8d:93", "network": {"id": "247a88e9-e8e8-46af-ac41-6777f0782fe0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1297422061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d93803f89ca64e66b8271e7b0bb48983", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddd93d49-eb", "ovs_interfaceid": "ddd93d49-eb57-4af9-a9bb-35d8269b939d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.825018] env[62952]: INFO nova.virt.block_device [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Booting with volume 22582603-2be6-47fd-860e-60d0d1fd0f86 at /dev/sda [ 1045.854068] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bec055b4-50b2-47ca-aa9a-03b5447fab4d tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.183s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.871643] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6eb3dd3f-9cee-429f-bfc0-71a1e32f91f5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.884253] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e7fcefa-8d00-435f-bd24-955cc7e0db06 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.921321] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-faba1a28-739e-4caf-a2e0-f08260826f9f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.928244] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3deb18af-d257-43c6-b6bb-a43854683f7d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.960303] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b37424d9-0335-4840-b8a4-6336e864dc62 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.968965] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ac1579-0b8d-453f-9202-0282ddd9c21a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.984799] env[62952]: DEBUG nova.virt.block_device [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Updating existing volume attachment record: 22dbb0ab-29b3-4e31-9538-629e66b67e11 {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1046.069378] env[62952]: DEBUG oslo_concurrency.lockutils [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.764s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.071631] env[62952]: DEBUG oslo_concurrency.lockutils [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.649s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.073234] env[62952]: INFO nova.compute.claims [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1046.086848] env[62952]: INFO nova.scheduler.client.report [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Deleted allocations for instance 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9 [ 1046.320275] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Releasing lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.320652] env[62952]: DEBUG nova.compute.manager [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Instance network_info: |[{"id": "ddd93d49-eb57-4af9-a9bb-35d8269b939d", "address": "fa:16:3e:e3:8d:93", "network": {"id": "247a88e9-e8e8-46af-ac41-6777f0782fe0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1297422061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d93803f89ca64e66b8271e7b0bb48983", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddd93d49-eb", "ovs_interfaceid": "ddd93d49-eb57-4af9-a9bb-35d8269b939d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1046.321252] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:8d:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4fcde7-8926-402a-a9b7-4878d2bc1cf6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ddd93d49-eb57-4af9-a9bb-35d8269b939d', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1046.329129] env[62952]: DEBUG oslo.service.loopingcall [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.329372] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1046.329596] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-097707af-5c45-480d-a5a4-882b8ebc20a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.349311] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1046.349311] env[62952]: value = "task-1263754" [ 1046.349311] env[62952]: _type = "Task" [ 1046.349311] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.357116] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263754, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.464862] env[62952]: DEBUG nova.compute.manager [req-77345961-4c5f-470d-a2b3-670f1542eab6 req-15ff2aef-2dae-48e4-ba77-8e6a312f960f service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Received event network-changed-ddd93d49-eb57-4af9-a9bb-35d8269b939d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1046.465234] env[62952]: DEBUG nova.compute.manager [req-77345961-4c5f-470d-a2b3-670f1542eab6 req-15ff2aef-2dae-48e4-ba77-8e6a312f960f service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Refreshing instance network info cache due to event network-changed-ddd93d49-eb57-4af9-a9bb-35d8269b939d. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1046.465578] env[62952]: DEBUG oslo_concurrency.lockutils [req-77345961-4c5f-470d-a2b3-670f1542eab6 req-15ff2aef-2dae-48e4-ba77-8e6a312f960f service nova] Acquiring lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.465808] env[62952]: DEBUG oslo_concurrency.lockutils [req-77345961-4c5f-470d-a2b3-670f1542eab6 req-15ff2aef-2dae-48e4-ba77-8e6a312f960f service nova] Acquired lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.466078] env[62952]: DEBUG nova.network.neutron [req-77345961-4c5f-470d-a2b3-670f1542eab6 req-15ff2aef-2dae-48e4-ba77-8e6a312f960f service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Refreshing network info cache for port ddd93d49-eb57-4af9-a9bb-35d8269b939d {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1046.511227] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1046.511461] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1046.511647] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Deleting the datastore file [datastore1] 05223053-90c3-433c-99b9-e252ef9a1935 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1046.511913] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-883f649b-b461-43c1-8904-c2bef3bf961d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.519037] env[62952]: DEBUG oslo_vmware.api [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1046.519037] env[62952]: value = "task-1263755" [ 1046.519037] env[62952]: _type = "Task" [ 1046.519037] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.527470] env[62952]: DEBUG oslo_vmware.api [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263755, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.595966] env[62952]: DEBUG oslo_concurrency.lockutils [None req-37d74de3-8a08-4f53-83df-3bb9a0f8537f tempest-InstanceActionsV221TestJSON-1564771547 tempest-InstanceActionsV221TestJSON-1564771547-project-member] Lock "4d83faf9-43db-4fc4-978b-55ca9a6ed4c9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.583s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.860315] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263754, 'name': CreateVM_Task, 'duration_secs': 0.38691} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.860592] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1046.861311] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.861560] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.861953] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1046.863887] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf94b6eb-5d47-4b55-a56c-0e1d64ae86a2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.873629] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1046.873629] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f1ebd1-1a82-d87e-4185-6d3ac5737129" [ 1046.873629] env[62952]: _type = "Task" [ 1046.873629] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.874103] env[62952]: DEBUG nova.compute.manager [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Stashing vm_state: active {{(pid=62952) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1046.889687] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f1ebd1-1a82-d87e-4185-6d3ac5737129, 'name': SearchDatastore_Task, 'duration_secs': 0.008617} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.889989] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.890274] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1046.890518] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.890669] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.890849] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1046.891129] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b84d9b95-024d-4ded-9a4d-2ef08274bd40 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.898393] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1046.898574] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1046.899323] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-307be032-368a-42b7-b87b-c9a15a181ffe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.904269] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1046.904269] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523e4da8-1178-0a67-355c-bd4ba02301ed" [ 1046.904269] env[62952]: _type = "Task" [ 1046.904269] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.911553] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523e4da8-1178-0a67-355c-bd4ba02301ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.029760] env[62952]: DEBUG oslo_vmware.api [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263755, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151515} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.030111] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1047.030355] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1047.030568] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1047.030924] env[62952]: INFO nova.compute.manager [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Took 2.12 seconds to destroy the instance on the hypervisor. [ 1047.031071] env[62952]: DEBUG oslo.service.loopingcall [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1047.033737] env[62952]: DEBUG nova.compute.manager [-] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1047.033858] env[62952]: DEBUG nova.network.neutron [-] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1047.299587] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6bf81f-ca2b-471c-b64a-b4dd8f123eea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.307528] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d69e18e-d714-4bcd-abca-ab2a528f7743 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.347785] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe41067e-b6c8-49e7-b84f-21e933422fa3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.358024] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19785b8d-7dc2-4011-878f-3042338bf76a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.370173] env[62952]: DEBUG nova.compute.provider_tree [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1047.401158] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.413730] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]523e4da8-1178-0a67-355c-bd4ba02301ed, 'name': SearchDatastore_Task, 'duration_secs': 0.00733} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.416686] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89f135c9-a041-41b5-96cb-d04c78f9e93a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.422363] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1047.422363] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5220a1b0-cb0f-b00c-084e-3c936119bdae" [ 1047.422363] env[62952]: _type = "Task" [ 1047.422363] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.432461] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5220a1b0-cb0f-b00c-084e-3c936119bdae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.499041] env[62952]: DEBUG nova.network.neutron [req-77345961-4c5f-470d-a2b3-670f1542eab6 req-15ff2aef-2dae-48e4-ba77-8e6a312f960f service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Updated VIF entry in instance network info cache for port ddd93d49-eb57-4af9-a9bb-35d8269b939d. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1047.499509] env[62952]: DEBUG nova.network.neutron [req-77345961-4c5f-470d-a2b3-670f1542eab6 req-15ff2aef-2dae-48e4-ba77-8e6a312f960f service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Updating instance_info_cache with network_info: [{"id": "ddd93d49-eb57-4af9-a9bb-35d8269b939d", "address": "fa:16:3e:e3:8d:93", "network": {"id": "247a88e9-e8e8-46af-ac41-6777f0782fe0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1297422061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d93803f89ca64e66b8271e7b0bb48983", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddd93d49-eb", "ovs_interfaceid": "ddd93d49-eb57-4af9-a9bb-35d8269b939d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.306572] env[62952]: DEBUG nova.network.neutron [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Successfully updated port: d5013042-c6e4-4118-aaa6-3947b89a6fb3 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1048.308461] env[62952]: DEBUG nova.scheduler.client.report [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1048.312213] env[62952]: DEBUG nova.network.neutron [-] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.313343] env[62952]: DEBUG oslo_concurrency.lockutils [req-77345961-4c5f-470d-a2b3-670f1542eab6 req-15ff2aef-2dae-48e4-ba77-8e6a312f960f service nova] Releasing lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.325621] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5220a1b0-cb0f-b00c-084e-3c936119bdae, 'name': SearchDatastore_Task, 'duration_secs': 0.009843} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.326233] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.326493] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 42793459-89e6-4941-9b20-cbe52a241c6d/42793459-89e6-4941-9b20-cbe52a241c6d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1048.327087] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6bc89bd2-f362-4502-9c87-198c3bba3fa7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.335047] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1048.335047] env[62952]: value = "task-1263756" [ 1048.335047] env[62952]: _type = "Task" [ 1048.335047] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.344141] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263756, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.492185] env[62952]: DEBUG nova.compute.manager [req-2e980423-bd60-482f-9569-e157d3923549 req-7c83a115-6bbd-4675-b97e-975dc7286cbe service nova] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Received event network-vif-plugged-d5013042-c6e4-4118-aaa6-3947b89a6fb3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1048.492357] env[62952]: DEBUG oslo_concurrency.lockutils [req-2e980423-bd60-482f-9569-e157d3923549 req-7c83a115-6bbd-4675-b97e-975dc7286cbe service nova] Acquiring lock "b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.492568] env[62952]: DEBUG oslo_concurrency.lockutils [req-2e980423-bd60-482f-9569-e157d3923549 req-7c83a115-6bbd-4675-b97e-975dc7286cbe service nova] Lock "b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.492770] env[62952]: DEBUG oslo_concurrency.lockutils [req-2e980423-bd60-482f-9569-e157d3923549 req-7c83a115-6bbd-4675-b97e-975dc7286cbe service nova] Lock "b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.492949] env[62952]: DEBUG nova.compute.manager [req-2e980423-bd60-482f-9569-e157d3923549 req-7c83a115-6bbd-4675-b97e-975dc7286cbe service nova] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] No waiting events found dispatching network-vif-plugged-d5013042-c6e4-4118-aaa6-3947b89a6fb3 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1048.493222] env[62952]: WARNING nova.compute.manager [req-2e980423-bd60-482f-9569-e157d3923549 req-7c83a115-6bbd-4675-b97e-975dc7286cbe service nova] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Received unexpected event network-vif-plugged-d5013042-c6e4-4118-aaa6-3947b89a6fb3 for instance with vm_state building and task_state spawning. [ 1048.493430] env[62952]: DEBUG nova.compute.manager [req-2e980423-bd60-482f-9569-e157d3923549 req-7c83a115-6bbd-4675-b97e-975dc7286cbe service nova] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Received event network-vif-deleted-fe560683-a241-423f-9f16-c5f59668d967 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1048.493604] env[62952]: DEBUG nova.compute.manager [req-2e980423-bd60-482f-9569-e157d3923549 req-7c83a115-6bbd-4675-b97e-975dc7286cbe service nova] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Received event network-changed-d5013042-c6e4-4118-aaa6-3947b89a6fb3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1048.493885] env[62952]: DEBUG nova.compute.manager [req-2e980423-bd60-482f-9569-e157d3923549 req-7c83a115-6bbd-4675-b97e-975dc7286cbe service nova] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Refreshing instance network info cache due to event network-changed-d5013042-c6e4-4118-aaa6-3947b89a6fb3. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1048.494328] env[62952]: DEBUG oslo_concurrency.lockutils [req-2e980423-bd60-482f-9569-e157d3923549 req-7c83a115-6bbd-4675-b97e-975dc7286cbe service nova] Acquiring lock "refresh_cache-b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.494328] env[62952]: DEBUG oslo_concurrency.lockutils [req-2e980423-bd60-482f-9569-e157d3923549 req-7c83a115-6bbd-4675-b97e-975dc7286cbe service nova] Acquired lock "refresh_cache-b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.494469] env[62952]: DEBUG nova.network.neutron [req-2e980423-bd60-482f-9569-e157d3923549 req-7c83a115-6bbd-4675-b97e-975dc7286cbe service nova] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Refreshing network info cache for port d5013042-c6e4-4118-aaa6-3947b89a6fb3 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1048.815951] env[62952]: DEBUG nova.compute.manager [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1048.816345] env[62952]: DEBUG nova.virt.hardware [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1048.816616] env[62952]: DEBUG nova.virt.hardware [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1048.816891] env[62952]: DEBUG nova.virt.hardware [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1048.817192] env[62952]: DEBUG nova.virt.hardware [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1048.817419] env[62952]: DEBUG nova.virt.hardware [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1048.817632] env[62952]: DEBUG nova.virt.hardware [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1048.817900] env[62952]: DEBUG nova.virt.hardware [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1048.818160] env[62952]: DEBUG nova.virt.hardware [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1048.818388] env[62952]: DEBUG nova.virt.hardware [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1048.818629] env[62952]: DEBUG nova.virt.hardware [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1048.818868] env[62952]: DEBUG nova.virt.hardware [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1048.819654] env[62952]: DEBUG oslo_concurrency.lockutils [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.748s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.820192] env[62952]: DEBUG nova.compute.manager [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1048.822807] env[62952]: DEBUG oslo_concurrency.lockutils [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Acquiring lock "refresh_cache-b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.823876] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9f11ee-0380-460c-acf4-3684f2460cf2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.826569] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.677s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.826784] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.826973] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1048.827298] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 5.181s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.827505] env[62952]: DEBUG nova.objects.instance [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62952) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1048.830280] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a88bc028-557a-44d3-99b5-cb4552b0d64e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.833579] env[62952]: INFO nova.compute.manager [-] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Took 1.80 seconds to deallocate network for instance. [ 1048.846016] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa7ecc86-c721-433f-9dc1-4536f477a129 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.852213] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-768a3205-d707-4837-83b6-f759aaca2642 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.882019] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263756, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.43721} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.883078] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 42793459-89e6-4941-9b20-cbe52a241c6d/42793459-89e6-4941-9b20-cbe52a241c6d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1048.883371] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1048.884181] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfa9d326-a8d1-4fe1-b861-53daefc38f0e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.887285] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5505e710-8717-4fc0-8777-ab76d462f6fa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.896282] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630be22d-a911-4e0d-89df-2dcc7c55fb2c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.899343] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1048.899343] env[62952]: value = "task-1263757" [ 1048.899343] env[62952]: _type = "Task" [ 1048.899343] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.926804] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180582MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1048.926985] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.931968] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263757, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.033098] env[62952]: DEBUG nova.network.neutron [req-2e980423-bd60-482f-9569-e157d3923549 req-7c83a115-6bbd-4675-b97e-975dc7286cbe service nova] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1049.124783] env[62952]: DEBUG nova.network.neutron [req-2e980423-bd60-482f-9569-e157d3923549 req-7c83a115-6bbd-4675-b97e-975dc7286cbe service nova] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.325395] env[62952]: DEBUG nova.compute.utils [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1049.326969] env[62952]: DEBUG nova.compute.manager [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1049.327234] env[62952]: DEBUG nova.network.neutron [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1049.352193] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.375858] env[62952]: DEBUG nova.policy [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7eadc05f193645e6b6e86e9cdbbbaa71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '076fb3c53420447caf047f8422435152', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1049.409706] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263757, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061782} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.409991] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1049.411151] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e55595-5f94-4377-9def-ea40f4bbfeb9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.433046] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 42793459-89e6-4941-9b20-cbe52a241c6d/42793459-89e6-4941-9b20-cbe52a241c6d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.433340] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbb348d1-bb8d-4af7-ae7b-7a08b32d5662 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.453620] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1049.453620] env[62952]: value = "task-1263758" [ 1049.453620] env[62952]: _type = "Task" [ 1049.453620] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.462121] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263758, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.628162] env[62952]: DEBUG oslo_concurrency.lockutils [req-2e980423-bd60-482f-9569-e157d3923549 req-7c83a115-6bbd-4675-b97e-975dc7286cbe service nova] Releasing lock "refresh_cache-b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.628770] env[62952]: DEBUG oslo_concurrency.lockutils [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Acquired lock "refresh_cache-b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.629061] env[62952]: DEBUG nova.network.neutron [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1049.731964] env[62952]: DEBUG nova.network.neutron [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Successfully created port: 70b1d946-ae76-4aba-9c14-ecc65d044630 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1049.831163] env[62952]: DEBUG nova.compute.manager [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1049.849047] env[62952]: DEBUG oslo_concurrency.lockutils [None req-43c8c1f7-646b-4931-98e5-68f1f97148fb tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.021s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.850876] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.159s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.851024] env[62952]: DEBUG nova.objects.instance [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lazy-loading 'resources' on Instance uuid d98a88aa-be5f-4254-aaba-c6fa03c9269a {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.964605] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263758, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.178279] env[62952]: DEBUG nova.network.neutron [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1050.395166] env[62952]: DEBUG nova.network.neutron [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Updating instance_info_cache with network_info: [{"id": "d5013042-c6e4-4118-aaa6-3947b89a6fb3", "address": "fa:16:3e:ac:f8:d7", "network": {"id": "bac9b3eb-e4d1-436c-b5d4-50db41886dfa", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1618686039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bff36f8c24b44ea8afc9094153087c5c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5013042-c6", "ovs_interfaceid": "d5013042-c6e4-4118-aaa6-3947b89a6fb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.465757] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263758, 'name': ReconfigVM_Task, 'duration_secs': 0.694296} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.469186] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 42793459-89e6-4941-9b20-cbe52a241c6d/42793459-89e6-4941-9b20-cbe52a241c6d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.470119] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0cecde4c-c85c-4f21-beb7-7166d0c73b3d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.478237] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1050.478237] env[62952]: value = "task-1263759" [ 1050.478237] env[62952]: _type = "Task" [ 1050.478237] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.490674] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263759, 'name': Rename_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.603033] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebfc1768-6753-41e5-a247-9d80a13a4382 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.611120] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a07c016-2e2a-45a6-9515-a6568dde2e73 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.643982] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9c95e7-05b3-4a82-96a5-e2541c2e17d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.652223] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ca22a7-d5c4-44e3-b9bd-d8977431f650 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.667371] env[62952]: DEBUG nova.compute.provider_tree [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1050.864683] env[62952]: DEBUG nova.compute.manager [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1050.895435] env[62952]: DEBUG nova.virt.hardware [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1050.896468] env[62952]: DEBUG nova.virt.hardware [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1050.896468] env[62952]: DEBUG nova.virt.hardware [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1050.896468] env[62952]: DEBUG nova.virt.hardware [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1050.896468] env[62952]: DEBUG nova.virt.hardware [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1050.896468] env[62952]: DEBUG nova.virt.hardware [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1050.896661] env[62952]: DEBUG nova.virt.hardware [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1050.897352] env[62952]: DEBUG nova.virt.hardware [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1050.897352] env[62952]: DEBUG nova.virt.hardware [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1050.897705] env[62952]: DEBUG nova.virt.hardware [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1050.897933] env[62952]: DEBUG nova.virt.hardware [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1050.899050] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e83506a7-c08e-4563-9509-4ed9c45b9052 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.901971] env[62952]: DEBUG oslo_concurrency.lockutils [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Releasing lock "refresh_cache-b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.902334] env[62952]: DEBUG nova.compute.manager [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Instance network_info: |[{"id": "d5013042-c6e4-4118-aaa6-3947b89a6fb3", "address": "fa:16:3e:ac:f8:d7", "network": {"id": "bac9b3eb-e4d1-436c-b5d4-50db41886dfa", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1618686039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bff36f8c24b44ea8afc9094153087c5c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5013042-c6", "ovs_interfaceid": "d5013042-c6e4-4118-aaa6-3947b89a6fb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1050.902740] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:f8:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5446413d-c3b0-4cd2-a962-62240db178ac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd5013042-c6e4-4118-aaa6-3947b89a6fb3', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1050.911168] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Creating folder: Project (bff36f8c24b44ea8afc9094153087c5c). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1050.911900] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-759a3a94-36f0-4468-89a8-eb7d69488e9e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.919335] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f2a73f4-4ce3-4488-8413-7a3471075a04 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.935381] env[62952]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1050.935519] env[62952]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62952) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1050.935885] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Folder already exists: Project (bff36f8c24b44ea8afc9094153087c5c). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1050.936095] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Creating folder: Instances. Parent ref: group-v271940. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1050.936333] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b0d68854-a8c0-49fa-9648-84fd5cae3591 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.940729] env[62952]: DEBUG oslo_concurrency.lockutils [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.940953] env[62952]: DEBUG oslo_concurrency.lockutils [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.948904] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Created folder: Instances in parent group-v271940. [ 1050.948904] env[62952]: DEBUG oslo.service.loopingcall [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.948904] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1050.948904] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d69aa25c-826f-4761-8296-50f603cf283b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.973340] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1050.973340] env[62952]: value = "task-1263762" [ 1050.973340] env[62952]: _type = "Task" [ 1050.973340] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.986054] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263762, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.991037] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263759, 'name': Rename_Task, 'duration_secs': 0.139005} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.991354] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1050.992280] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f9c6c602-fa66-4801-808c-0f9ac1e10ba6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.003305] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1051.003305] env[62952]: value = "task-1263763" [ 1051.003305] env[62952]: _type = "Task" [ 1051.003305] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.011639] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263763, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.170636] env[62952]: DEBUG nova.scheduler.client.report [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1051.377488] env[62952]: DEBUG nova.compute.manager [req-fc5993a1-74cd-4ff8-be23-1a60267e6b8c req-848a81c0-59d1-4e8a-9a21-c246dba103b9 service nova] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Received event network-vif-plugged-70b1d946-ae76-4aba-9c14-ecc65d044630 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1051.377839] env[62952]: DEBUG oslo_concurrency.lockutils [req-fc5993a1-74cd-4ff8-be23-1a60267e6b8c req-848a81c0-59d1-4e8a-9a21-c246dba103b9 service nova] Acquiring lock "1a38bdcd-5537-46f5-94f6-a8e358e13121-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.378225] env[62952]: DEBUG oslo_concurrency.lockutils [req-fc5993a1-74cd-4ff8-be23-1a60267e6b8c req-848a81c0-59d1-4e8a-9a21-c246dba103b9 service nova] Lock "1a38bdcd-5537-46f5-94f6-a8e358e13121-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.378515] env[62952]: DEBUG oslo_concurrency.lockutils [req-fc5993a1-74cd-4ff8-be23-1a60267e6b8c req-848a81c0-59d1-4e8a-9a21-c246dba103b9 service nova] Lock "1a38bdcd-5537-46f5-94f6-a8e358e13121-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.378801] env[62952]: DEBUG nova.compute.manager [req-fc5993a1-74cd-4ff8-be23-1a60267e6b8c req-848a81c0-59d1-4e8a-9a21-c246dba103b9 service nova] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] No waiting events found dispatching network-vif-plugged-70b1d946-ae76-4aba-9c14-ecc65d044630 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1051.379106] env[62952]: WARNING nova.compute.manager [req-fc5993a1-74cd-4ff8-be23-1a60267e6b8c req-848a81c0-59d1-4e8a-9a21-c246dba103b9 service nova] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Received unexpected event network-vif-plugged-70b1d946-ae76-4aba-9c14-ecc65d044630 for instance with vm_state building and task_state spawning. [ 1051.436228] env[62952]: DEBUG nova.network.neutron [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Successfully updated port: 70b1d946-ae76-4aba-9c14-ecc65d044630 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1051.444818] env[62952]: DEBUG nova.compute.utils [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1051.488787] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263762, 'name': CreateVM_Task, 'duration_secs': 0.365293} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.489068] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1051.489888] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271949', 'volume_id': '22582603-2be6-47fd-860e-60d0d1fd0f86', 'name': 'volume-22582603-2be6-47fd-860e-60d0d1fd0f86', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b', 'attached_at': '', 'detached_at': '', 'volume_id': '22582603-2be6-47fd-860e-60d0d1fd0f86', 'serial': '22582603-2be6-47fd-860e-60d0d1fd0f86'}, 'boot_index': 0, 'mount_device': '/dev/sda', 'device_type': None, 'delete_on_termination': True, 'attachment_id': '22dbb0ab-29b3-4e31-9538-629e66b67e11', 'disk_bus': None, 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=62952) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1051.490189] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Root volume attach. Driver type: vmdk {{(pid=62952) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1051.491361] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77487cbe-d82a-44af-be43-2e168c0bcaab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.502427] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7abecc22-fc1d-44e5-adae-56bbc2d19171 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.515118] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263763, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.517116] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882c0dab-d2b1-443b-aed0-3c7a9db50c89 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.523163] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-b5aa62ab-e52b-43e2-8253-be9fc68124b9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.531507] env[62952]: DEBUG oslo_vmware.api [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Waiting for the task: (returnval){ [ 1051.531507] env[62952]: value = "task-1263764" [ 1051.531507] env[62952]: _type = "Task" [ 1051.531507] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.539365] env[62952]: DEBUG oslo_vmware.api [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263764, 'name': RelocateVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.676334] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.825s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.679168] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.457s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.681348] env[62952]: INFO nova.compute.claims [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1051.701619] env[62952]: INFO nova.scheduler.client.report [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Deleted allocations for instance d98a88aa-be5f-4254-aaba-c6fa03c9269a [ 1051.762252] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Acquiring lock "1a02fba4-aad0-4bd4-8183-a7f08d887a5a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.762509] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Lock "1a02fba4-aad0-4bd4-8183-a7f08d887a5a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.937472] env[62952]: DEBUG oslo_concurrency.lockutils [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "refresh_cache-1a38bdcd-5537-46f5-94f6-a8e358e13121" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.937630] env[62952]: DEBUG oslo_concurrency.lockutils [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "refresh_cache-1a38bdcd-5537-46f5-94f6-a8e358e13121" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.937790] env[62952]: DEBUG nova.network.neutron [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1051.950156] env[62952]: DEBUG oslo_concurrency.lockutils [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.017640] env[62952]: DEBUG oslo_vmware.api [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263763, 'name': PowerOnVM_Task, 'duration_secs': 0.722087} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.017640] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1052.017640] env[62952]: INFO nova.compute.manager [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Took 7.98 seconds to spawn the instance on the hypervisor. [ 1052.017640] env[62952]: DEBUG nova.compute.manager [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1052.017896] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e64d168-41ab-4c8e-9797-db0c7ed885f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.041030] env[62952]: DEBUG oslo_vmware.api [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263764, 'name': RelocateVM_Task, 'duration_secs': 0.416059} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.041030] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1052.041254] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271949', 'volume_id': '22582603-2be6-47fd-860e-60d0d1fd0f86', 'name': 'volume-22582603-2be6-47fd-860e-60d0d1fd0f86', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b', 'attached_at': '', 'detached_at': '', 'volume_id': '22582603-2be6-47fd-860e-60d0d1fd0f86', 'serial': '22582603-2be6-47fd-860e-60d0d1fd0f86'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1052.045020] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6774381d-7360-4739-9ebd-9ee7c1663f74 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.058141] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9dbf73-5820-4fe8-95bc-2a4d3694949b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.082999] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] volume-22582603-2be6-47fd-860e-60d0d1fd0f86/volume-22582603-2be6-47fd-860e-60d0d1fd0f86.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1052.083599] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a54c69d3-db9c-42a4-a887-8cb5c0e4744d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.103725] env[62952]: DEBUG oslo_vmware.api [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Waiting for the task: (returnval){ [ 1052.103725] env[62952]: value = "task-1263765" [ 1052.103725] env[62952]: _type = "Task" [ 1052.103725] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.111824] env[62952]: DEBUG oslo_vmware.api [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263765, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.213029] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fb3585fc-f676-4fc7-95a9-7a7c64bc830e tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "d98a88aa-be5f-4254-aaba-c6fa03c9269a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.410s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.265155] env[62952]: DEBUG nova.compute.manager [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1052.489466] env[62952]: DEBUG nova.network.neutron [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1052.540757] env[62952]: INFO nova.compute.manager [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Took 18.33 seconds to build instance. [ 1052.613668] env[62952]: DEBUG oslo_vmware.api [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263765, 'name': ReconfigVM_Task, 'duration_secs': 0.446431} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.614012] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Reconfigured VM instance instance-0000005f to attach disk [datastore2] volume-22582603-2be6-47fd-860e-60d0d1fd0f86/volume-22582603-2be6-47fd-860e-60d0d1fd0f86.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1052.618898] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9898ad3-7568-42a3-82a2-a299f5a839a6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.646281] env[62952]: DEBUG oslo_vmware.api [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Waiting for the task: (returnval){ [ 1052.646281] env[62952]: value = "task-1263766" [ 1052.646281] env[62952]: _type = "Task" [ 1052.646281] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.656910] env[62952]: DEBUG oslo_vmware.api [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263766, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.694296] env[62952]: DEBUG nova.network.neutron [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Updating instance_info_cache with network_info: [{"id": "70b1d946-ae76-4aba-9c14-ecc65d044630", "address": "fa:16:3e:aa:53:4a", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70b1d946-ae", "ovs_interfaceid": "70b1d946-ae76-4aba-9c14-ecc65d044630", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.799019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.938197] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c77a84c7-148f-4895-a3f9-698389249d63 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.945920] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4d1d37-8d36-4a8d-9912-ed347c327b9a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.978142] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62d88b0-ca0b-474e-966d-6d20d5ebb0e8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.985604] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01109173-4985-47e0-8ad3-133277368d48 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.998764] env[62952]: DEBUG nova.compute.provider_tree [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.034010] env[62952]: DEBUG oslo_concurrency.lockutils [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.034283] env[62952]: DEBUG oslo_concurrency.lockutils [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.034718] env[62952]: INFO nova.compute.manager [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Attaching volume 3b174d32-3a39-4759-ba6c-6ea837f89b2e to /dev/sdb [ 1053.043478] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50849864-08b8-41dc-b8c5-d14696ae38a5 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "42793459-89e6-4941-9b20-cbe52a241c6d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.846s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.076825] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5efd08e-3177-40d4-afe1-1971b52ffe1d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.084399] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-447f6764-386b-4628-baea-77f9cf9135b4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.090112] env[62952]: DEBUG nova.compute.manager [req-ec1e5d7e-25c8-4bad-b32d-8041dc2478c5 req-6fffbf4e-0222-471d-a7aa-6546dac24fc0 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Received event network-changed-ddd93d49-eb57-4af9-a9bb-35d8269b939d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1053.090317] env[62952]: DEBUG nova.compute.manager [req-ec1e5d7e-25c8-4bad-b32d-8041dc2478c5 req-6fffbf4e-0222-471d-a7aa-6546dac24fc0 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Refreshing instance network info cache due to event network-changed-ddd93d49-eb57-4af9-a9bb-35d8269b939d. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1053.090562] env[62952]: DEBUG oslo_concurrency.lockutils [req-ec1e5d7e-25c8-4bad-b32d-8041dc2478c5 req-6fffbf4e-0222-471d-a7aa-6546dac24fc0 service nova] Acquiring lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.090664] env[62952]: DEBUG oslo_concurrency.lockutils [req-ec1e5d7e-25c8-4bad-b32d-8041dc2478c5 req-6fffbf4e-0222-471d-a7aa-6546dac24fc0 service nova] Acquired lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.090821] env[62952]: DEBUG nova.network.neutron [req-ec1e5d7e-25c8-4bad-b32d-8041dc2478c5 req-6fffbf4e-0222-471d-a7aa-6546dac24fc0 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Refreshing network info cache for port ddd93d49-eb57-4af9-a9bb-35d8269b939d {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1053.097421] env[62952]: DEBUG nova.virt.block_device [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Updating existing volume attachment record: 2ff0bcd1-3e98-4f1a-8276-11c884b083ec {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1053.141950] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "7ba1587d-edaf-4dce-a224-6b2ac22cfecd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.142242] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "7ba1587d-edaf-4dce-a224-6b2ac22cfecd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.142456] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "7ba1587d-edaf-4dce-a224-6b2ac22cfecd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.142641] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "7ba1587d-edaf-4dce-a224-6b2ac22cfecd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.142815] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "7ba1587d-edaf-4dce-a224-6b2ac22cfecd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.145866] env[62952]: INFO nova.compute.manager [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Terminating instance [ 1053.151459] env[62952]: DEBUG nova.compute.manager [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1053.151651] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1053.152401] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c1d5aa-f211-4e97-8db5-d5a52aa15d85 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.159677] env[62952]: DEBUG oslo_vmware.api [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263766, 'name': ReconfigVM_Task, 'duration_secs': 0.12606} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.161641] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271949', 'volume_id': '22582603-2be6-47fd-860e-60d0d1fd0f86', 'name': 'volume-22582603-2be6-47fd-860e-60d0d1fd0f86', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b', 'attached_at': '', 'detached_at': '', 'volume_id': '22582603-2be6-47fd-860e-60d0d1fd0f86', 'serial': '22582603-2be6-47fd-860e-60d0d1fd0f86'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1053.162189] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1053.162405] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7ab11949-83fb-4678-8436-05460cfbeca1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.163801] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cad8a71a-202d-4ace-ba01-52c77d6b8044 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.169283] env[62952]: DEBUG oslo_vmware.api [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Waiting for the task: (returnval){ [ 1053.169283] env[62952]: value = "task-1263767" [ 1053.169283] env[62952]: _type = "Task" [ 1053.169283] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.172996] env[62952]: DEBUG oslo_vmware.api [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1053.172996] env[62952]: value = "task-1263768" [ 1053.172996] env[62952]: _type = "Task" [ 1053.172996] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.180236] env[62952]: DEBUG oslo_vmware.api [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263767, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.185471] env[62952]: DEBUG oslo_vmware.api [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263768, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.198187] env[62952]: DEBUG oslo_concurrency.lockutils [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "refresh_cache-1a38bdcd-5537-46f5-94f6-a8e358e13121" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.198683] env[62952]: DEBUG nova.compute.manager [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Instance network_info: |[{"id": "70b1d946-ae76-4aba-9c14-ecc65d044630", "address": "fa:16:3e:aa:53:4a", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70b1d946-ae", "ovs_interfaceid": "70b1d946-ae76-4aba-9c14-ecc65d044630", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1053.199285] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:aa:53:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '365ac5b1-6d83-4dfe-887f-60574d7f6124', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '70b1d946-ae76-4aba-9c14-ecc65d044630', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1053.207188] env[62952]: DEBUG oslo.service.loopingcall [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.207431] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1053.207676] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-046c650a-3c1b-4d7c-93ee-a9667bc42ae8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.227116] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1053.227116] env[62952]: value = "task-1263769" [ 1053.227116] env[62952]: _type = "Task" [ 1053.227116] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.239155] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263769, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.405067] env[62952]: DEBUG nova.compute.manager [req-ff0eb488-23b9-4ff6-b90a-af7f253b0615 req-9b1d68d4-f7eb-4663-a129-6de3617645b2 service nova] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Received event network-changed-70b1d946-ae76-4aba-9c14-ecc65d044630 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1053.405685] env[62952]: DEBUG nova.compute.manager [req-ff0eb488-23b9-4ff6-b90a-af7f253b0615 req-9b1d68d4-f7eb-4663-a129-6de3617645b2 service nova] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Refreshing instance network info cache due to event network-changed-70b1d946-ae76-4aba-9c14-ecc65d044630. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1053.405993] env[62952]: DEBUG oslo_concurrency.lockutils [req-ff0eb488-23b9-4ff6-b90a-af7f253b0615 req-9b1d68d4-f7eb-4663-a129-6de3617645b2 service nova] Acquiring lock "refresh_cache-1a38bdcd-5537-46f5-94f6-a8e358e13121" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.406234] env[62952]: DEBUG oslo_concurrency.lockutils [req-ff0eb488-23b9-4ff6-b90a-af7f253b0615 req-9b1d68d4-f7eb-4663-a129-6de3617645b2 service nova] Acquired lock "refresh_cache-1a38bdcd-5537-46f5-94f6-a8e358e13121" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.406459] env[62952]: DEBUG nova.network.neutron [req-ff0eb488-23b9-4ff6-b90a-af7f253b0615 req-9b1d68d4-f7eb-4663-a129-6de3617645b2 service nova] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Refreshing network info cache for port 70b1d946-ae76-4aba-9c14-ecc65d044630 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1053.501703] env[62952]: DEBUG nova.scheduler.client.report [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1053.684522] env[62952]: DEBUG oslo_vmware.api [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263768, 'name': PowerOffVM_Task, 'duration_secs': 0.230875} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.684778] env[62952]: DEBUG oslo_vmware.api [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263767, 'name': Rename_Task, 'duration_secs': 0.130949} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.685017] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1053.685193] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1053.685485] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1053.685724] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3a974186-c76a-4837-93e2-dd66662c759e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.687141] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a86cde2-1fe6-43d3-b545-556e9018f758 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.694708] env[62952]: DEBUG oslo_vmware.api [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Waiting for the task: (returnval){ [ 1053.694708] env[62952]: value = "task-1263772" [ 1053.694708] env[62952]: _type = "Task" [ 1053.694708] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.702559] env[62952]: DEBUG oslo_vmware.api [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263772, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.736236] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263769, 'name': CreateVM_Task, 'duration_secs': 0.333019} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.736392] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1053.737096] env[62952]: DEBUG oslo_concurrency.lockutils [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.737632] env[62952]: DEBUG oslo_concurrency.lockutils [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.737773] env[62952]: DEBUG oslo_concurrency.lockutils [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1053.738139] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1097fa5c-950e-4fe3-b8ff-1fdc5bb9a19d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.742910] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1053.742910] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c4b4f6-2062-43ba-47a8-141d8f198a1f" [ 1053.742910] env[62952]: _type = "Task" [ 1053.742910] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.750369] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c4b4f6-2062-43ba-47a8-141d8f198a1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.820018] env[62952]: DEBUG nova.network.neutron [req-ec1e5d7e-25c8-4bad-b32d-8041dc2478c5 req-6fffbf4e-0222-471d-a7aa-6546dac24fc0 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Updated VIF entry in instance network info cache for port ddd93d49-eb57-4af9-a9bb-35d8269b939d. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1053.820524] env[62952]: DEBUG nova.network.neutron [req-ec1e5d7e-25c8-4bad-b32d-8041dc2478c5 req-6fffbf4e-0222-471d-a7aa-6546dac24fc0 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Updating instance_info_cache with network_info: [{"id": "ddd93d49-eb57-4af9-a9bb-35d8269b939d", "address": "fa:16:3e:e3:8d:93", "network": {"id": "247a88e9-e8e8-46af-ac41-6777f0782fe0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1297422061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.226", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d93803f89ca64e66b8271e7b0bb48983", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddd93d49-eb", "ovs_interfaceid": "ddd93d49-eb57-4af9-a9bb-35d8269b939d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.905290] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1053.905630] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1053.905869] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Deleting the datastore file [datastore2] 7ba1587d-edaf-4dce-a224-6b2ac22cfecd {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1053.906198] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f5a8029-2f7c-4b6c-ada5-93e0488bdcf1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.912958] env[62952]: DEBUG oslo_vmware.api [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for the task: (returnval){ [ 1053.912958] env[62952]: value = "task-1263773" [ 1053.912958] env[62952]: _type = "Task" [ 1053.912958] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.920910] env[62952]: DEBUG oslo_vmware.api [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263773, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.007111] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.328s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.007374] env[62952]: DEBUG nova.compute.manager [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1054.011020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.609s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.206774] env[62952]: DEBUG oslo_vmware.api [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263772, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.253892] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c4b4f6-2062-43ba-47a8-141d8f198a1f, 'name': SearchDatastore_Task, 'duration_secs': 0.012513} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.257018] env[62952]: DEBUG oslo_concurrency.lockutils [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.257018] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1054.257018] env[62952]: DEBUG oslo_concurrency.lockutils [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.257018] env[62952]: DEBUG oslo_concurrency.lockutils [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.257018] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1054.257018] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-387fb2c9-1567-4238-b56d-2961d2c46647 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.264510] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1054.264922] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1054.265780] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f517e13f-9cd7-4a6e-b46b-9bcc91542b59 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.273172] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1054.273172] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5258c6b9-2f91-a300-2f4a-1325970e8710" [ 1054.273172] env[62952]: _type = "Task" [ 1054.273172] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.279025] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5258c6b9-2f91-a300-2f4a-1325970e8710, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.283911] env[62952]: DEBUG nova.network.neutron [req-ff0eb488-23b9-4ff6-b90a-af7f253b0615 req-9b1d68d4-f7eb-4663-a129-6de3617645b2 service nova] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Updated VIF entry in instance network info cache for port 70b1d946-ae76-4aba-9c14-ecc65d044630. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1054.284283] env[62952]: DEBUG nova.network.neutron [req-ff0eb488-23b9-4ff6-b90a-af7f253b0615 req-9b1d68d4-f7eb-4663-a129-6de3617645b2 service nova] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Updating instance_info_cache with network_info: [{"id": "70b1d946-ae76-4aba-9c14-ecc65d044630", "address": "fa:16:3e:aa:53:4a", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70b1d946-ae", "ovs_interfaceid": "70b1d946-ae76-4aba-9c14-ecc65d044630", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.323733] env[62952]: DEBUG oslo_concurrency.lockutils [req-ec1e5d7e-25c8-4bad-b32d-8041dc2478c5 req-6fffbf4e-0222-471d-a7aa-6546dac24fc0 service nova] Releasing lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.423882] env[62952]: DEBUG oslo_vmware.api [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263773, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.514522] env[62952]: DEBUG nova.compute.utils [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1054.516067] env[62952]: DEBUG nova.compute.manager [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1054.516251] env[62952]: DEBUG nova.network.neutron [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1054.521275] env[62952]: INFO nova.compute.claims [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1054.564516] env[62952]: DEBUG nova.policy [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '469eae9db3d14e9f90988e3093514e69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff928791a3cf496cb8db2c2a6c47a6a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1054.704064] env[62952]: DEBUG oslo_vmware.api [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263772, 'name': PowerOnVM_Task, 'duration_secs': 0.881198} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.704236] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1054.704434] env[62952]: INFO nova.compute.manager [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Took 5.89 seconds to spawn the instance on the hypervisor. [ 1054.704616] env[62952]: DEBUG nova.compute.manager [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1054.705422] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34b3e7a-6732-43f3-bc23-992056631e8c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.782324] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5258c6b9-2f91-a300-2f4a-1325970e8710, 'name': SearchDatastore_Task, 'duration_secs': 0.008212} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.783064] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2567b3f8-716b-441c-b26d-9cf0ad575e2c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.786973] env[62952]: DEBUG oslo_concurrency.lockutils [req-ff0eb488-23b9-4ff6-b90a-af7f253b0615 req-9b1d68d4-f7eb-4663-a129-6de3617645b2 service nova] Releasing lock "refresh_cache-1a38bdcd-5537-46f5-94f6-a8e358e13121" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.788482] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1054.788482] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52decb79-8f03-2856-dbc6-4a3d1e2d9bf9" [ 1054.788482] env[62952]: _type = "Task" [ 1054.788482] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.795800] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52decb79-8f03-2856-dbc6-4a3d1e2d9bf9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.836540] env[62952]: DEBUG nova.network.neutron [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Successfully created port: 3ab932ab-2d08-4a0f-af17-e69bfb046602 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1054.923637] env[62952]: DEBUG oslo_vmware.api [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Task: {'id': task-1263773, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.535672} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.923895] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1054.924101] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1054.924284] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1054.924464] env[62952]: INFO nova.compute.manager [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Took 1.77 seconds to destroy the instance on the hypervisor. [ 1054.924699] env[62952]: DEBUG oslo.service.loopingcall [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1054.924891] env[62952]: DEBUG nova.compute.manager [-] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1054.924985] env[62952]: DEBUG nova.network.neutron [-] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1055.025636] env[62952]: DEBUG nova.compute.manager [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1055.029812] env[62952]: INFO nova.compute.resource_tracker [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Updating resource usage from migration fdd5dcdb-d6fb-4c12-a4dd-cbd88d9a7609 [ 1055.222906] env[62952]: INFO nova.compute.manager [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Took 20.02 seconds to build instance. [ 1055.248879] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0860a48-9d9f-46cd-998b-1618f7f189b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.257266] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-952a999c-dbdf-4113-8847-3a3cc9cc156a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.299132] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1495e532-c17f-4ef1-882f-6ea699fd8d75 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.304608] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52decb79-8f03-2856-dbc6-4a3d1e2d9bf9, 'name': SearchDatastore_Task, 'duration_secs': 0.008838} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.306756] env[62952]: DEBUG oslo_concurrency.lockutils [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.307034] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 1a38bdcd-5537-46f5-94f6-a8e358e13121/1a38bdcd-5537-46f5-94f6-a8e358e13121.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1055.307338] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5537d8c8-60db-4f78-93a3-99216e2805ff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.310132] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e91d293-c6c0-4af7-a23e-5ba8f596cf5d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.325367] env[62952]: DEBUG nova.compute.provider_tree [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.328612] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1055.328612] env[62952]: value = "task-1263775" [ 1055.328612] env[62952]: _type = "Task" [ 1055.328612] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.336666] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263775, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.437493] env[62952]: DEBUG nova.compute.manager [req-977aaa23-4561-4270-9bcf-4150ebb3d759 req-5d3c652d-5d9b-483c-a4e7-96c85b657e1b service nova] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Received event network-vif-deleted-a6b1b43b-91ae-45fe-8e98-bd3a194bd05b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1055.437493] env[62952]: INFO nova.compute.manager [req-977aaa23-4561-4270-9bcf-4150ebb3d759 req-5d3c652d-5d9b-483c-a4e7-96c85b657e1b service nova] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Neutron deleted interface a6b1b43b-91ae-45fe-8e98-bd3a194bd05b; detaching it from the instance and deleting it from the info cache [ 1055.437493] env[62952]: DEBUG nova.network.neutron [req-977aaa23-4561-4270-9bcf-4150ebb3d759 req-5d3c652d-5d9b-483c-a4e7-96c85b657e1b service nova] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.661847] env[62952]: DEBUG nova.network.neutron [-] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.724635] env[62952]: DEBUG oslo_concurrency.lockutils [None req-00e91442-8f7c-4a09-914c-8f359f6f0feb tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Lock "b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.539s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.828737] env[62952]: DEBUG nova.scheduler.client.report [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1055.843781] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263775, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44642} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.844064] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 1a38bdcd-5537-46f5-94f6-a8e358e13121/1a38bdcd-5537-46f5-94f6-a8e358e13121.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1055.844309] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1055.844627] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2b772d0b-25a3-4527-b84a-b7c867e2bbcb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.852681] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1055.852681] env[62952]: value = "task-1263776" [ 1055.852681] env[62952]: _type = "Task" [ 1055.852681] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.861940] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263776, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.940234] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b598b5b-e1e5-4984-9d26-c20b7f9d2693 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.948740] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-118a7110-460e-4fd9-b5c9-c116f018ac12 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.976797] env[62952]: DEBUG nova.compute.manager [req-977aaa23-4561-4270-9bcf-4150ebb3d759 req-5d3c652d-5d9b-483c-a4e7-96c85b657e1b service nova] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Detach interface failed, port_id=a6b1b43b-91ae-45fe-8e98-bd3a194bd05b, reason: Instance 7ba1587d-edaf-4dce-a224-6b2ac22cfecd could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1056.038851] env[62952]: DEBUG nova.compute.manager [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1056.066598] env[62952]: DEBUG nova.virt.hardware [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1056.066866] env[62952]: DEBUG nova.virt.hardware [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1056.067040] env[62952]: DEBUG nova.virt.hardware [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1056.067242] env[62952]: DEBUG nova.virt.hardware [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1056.067402] env[62952]: DEBUG nova.virt.hardware [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1056.067557] env[62952]: DEBUG nova.virt.hardware [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1056.067801] env[62952]: DEBUG nova.virt.hardware [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1056.067991] env[62952]: DEBUG nova.virt.hardware [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1056.068177] env[62952]: DEBUG nova.virt.hardware [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1056.068348] env[62952]: DEBUG nova.virt.hardware [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1056.068525] env[62952]: DEBUG nova.virt.hardware [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1056.069512] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa0d5bc-4f31-4e70-ad2c-9bae46b868eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.077300] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf350ac-3b5f-4151-a4ea-f52f5d3fa811 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.164612] env[62952]: INFO nova.compute.manager [-] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Took 1.24 seconds to deallocate network for instance. [ 1056.337377] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.327s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.337601] env[62952]: INFO nova.compute.manager [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Migrating [ 1056.344254] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 7.417s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.362120] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263776, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061683} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.362335] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1056.363125] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e205d1b-0e49-4a04-af19-e03c5fe55395 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.385702] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 1a38bdcd-5537-46f5-94f6-a8e358e13121/1a38bdcd-5537-46f5-94f6-a8e358e13121.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1056.386274] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31be3843-919e-41b4-ae05-1ccca044e5ed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.409795] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1056.409795] env[62952]: value = "task-1263777" [ 1056.409795] env[62952]: _type = "Task" [ 1056.409795] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.417972] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263777, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.571325] env[62952]: DEBUG nova.compute.manager [req-4197bec9-00dc-4d53-813c-36ed1c333f8b req-a98e3333-2112-497e-9703-c0f78ef6f590 service nova] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Received event network-vif-plugged-3ab932ab-2d08-4a0f-af17-e69bfb046602 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1056.571325] env[62952]: DEBUG oslo_concurrency.lockutils [req-4197bec9-00dc-4d53-813c-36ed1c333f8b req-a98e3333-2112-497e-9703-c0f78ef6f590 service nova] Acquiring lock "af261b32-457a-44a1-bab4-b62654b0be1f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.571325] env[62952]: DEBUG oslo_concurrency.lockutils [req-4197bec9-00dc-4d53-813c-36ed1c333f8b req-a98e3333-2112-497e-9703-c0f78ef6f590 service nova] Lock "af261b32-457a-44a1-bab4-b62654b0be1f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.571325] env[62952]: DEBUG oslo_concurrency.lockutils [req-4197bec9-00dc-4d53-813c-36ed1c333f8b req-a98e3333-2112-497e-9703-c0f78ef6f590 service nova] Lock "af261b32-457a-44a1-bab4-b62654b0be1f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.571325] env[62952]: DEBUG nova.compute.manager [req-4197bec9-00dc-4d53-813c-36ed1c333f8b req-a98e3333-2112-497e-9703-c0f78ef6f590 service nova] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] No waiting events found dispatching network-vif-plugged-3ab932ab-2d08-4a0f-af17-e69bfb046602 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1056.572167] env[62952]: WARNING nova.compute.manager [req-4197bec9-00dc-4d53-813c-36ed1c333f8b req-a98e3333-2112-497e-9703-c0f78ef6f590 service nova] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Received unexpected event network-vif-plugged-3ab932ab-2d08-4a0f-af17-e69bfb046602 for instance with vm_state building and task_state spawning. [ 1056.671894] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.797037] env[62952]: DEBUG nova.network.neutron [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Successfully updated port: 3ab932ab-2d08-4a0f-af17-e69bfb046602 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1056.857409] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "refresh_cache-e9e690b8-2fae-401d-aaad-0acef6e88ce3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.857612] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "refresh_cache-e9e690b8-2fae-401d-aaad-0acef6e88ce3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.857737] env[62952]: DEBUG nova.network.neutron [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1056.921115] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263777, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.298549] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "refresh_cache-af261b32-457a-44a1-bab4-b62654b0be1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.298776] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "refresh_cache-af261b32-457a-44a1-bab4-b62654b0be1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.298981] env[62952]: DEBUG nova.network.neutron [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1057.360819] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Applying migration context for instance e9e690b8-2fae-401d-aaad-0acef6e88ce3 as it has an incoming, in-progress migration fdd5dcdb-d6fb-4c12-a4dd-cbd88d9a7609. Migration status is pre-migrating {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1057.362363] env[62952]: INFO nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Updating resource usage from migration fdd5dcdb-d6fb-4c12-a4dd-cbd88d9a7609 [ 1057.384924] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 6e289c20-8dd2-4680-b816-9bbe82fe7aec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1057.385099] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance dbca4369-1ed3-493a-9847-9fa1e4293475 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1057.385226] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 77fe79de-8da8-486e-af63-91f9d8196ca9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1057.385375] env[62952]: WARNING nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 7ba1587d-edaf-4dce-a224-6b2ac22cfecd is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1057.385486] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance e44914cd-d8ae-4f43-9207-32cacf6d6553 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1057.385609] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 1941365e-d562-4d99-89eb-226a07e52071 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1057.385769] env[62952]: WARNING nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 05223053-90c3-433c-99b9-e252ef9a1935 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1057.385890] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 42793459-89e6-4941-9b20-cbe52a241c6d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1057.386010] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1057.386132] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 1a38bdcd-5537-46f5-94f6-a8e358e13121 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1057.386243] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance af261b32-457a-44a1-bab4-b62654b0be1f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1057.386354] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Migration fdd5dcdb-d6fb-4c12-a4dd-cbd88d9a7609 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1057.386508] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance e9e690b8-2fae-401d-aaad-0acef6e88ce3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1057.420764] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263777, 'name': ReconfigVM_Task, 'duration_secs': 0.723378} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.420929] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 1a38bdcd-5537-46f5-94f6-a8e358e13121/1a38bdcd-5537-46f5-94f6-a8e358e13121.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1057.421547] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ad8d7cd9-9329-47cd-868d-b40b7d6e1a0b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.427559] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1057.427559] env[62952]: value = "task-1263778" [ 1057.427559] env[62952]: _type = "Task" [ 1057.427559] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.435593] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263778, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.462248] env[62952]: DEBUG nova.compute.manager [req-486d9796-c15e-4246-88a7-701872487aac req-81d42f12-0ab4-482a-8f9d-e2085caa38ca service nova] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Received event network-changed-d5013042-c6e4-4118-aaa6-3947b89a6fb3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1057.462454] env[62952]: DEBUG nova.compute.manager [req-486d9796-c15e-4246-88a7-701872487aac req-81d42f12-0ab4-482a-8f9d-e2085caa38ca service nova] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Refreshing instance network info cache due to event network-changed-d5013042-c6e4-4118-aaa6-3947b89a6fb3. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1057.462680] env[62952]: DEBUG oslo_concurrency.lockutils [req-486d9796-c15e-4246-88a7-701872487aac req-81d42f12-0ab4-482a-8f9d-e2085caa38ca service nova] Acquiring lock "refresh_cache-b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.462810] env[62952]: DEBUG oslo_concurrency.lockutils [req-486d9796-c15e-4246-88a7-701872487aac req-81d42f12-0ab4-482a-8f9d-e2085caa38ca service nova] Acquired lock "refresh_cache-b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.462975] env[62952]: DEBUG nova.network.neutron [req-486d9796-c15e-4246-88a7-701872487aac req-81d42f12-0ab4-482a-8f9d-e2085caa38ca service nova] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Refreshing network info cache for port d5013042-c6e4-4118-aaa6-3947b89a6fb3 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1057.576536] env[62952]: DEBUG nova.network.neutron [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Updating instance_info_cache with network_info: [{"id": "cff4c81c-513a-42a4-b6f5-94b4367fa0ea", "address": "fa:16:3e:3a:e4:22", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcff4c81c-51", "ovs_interfaceid": "cff4c81c-513a-42a4-b6f5-94b4367fa0ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.642109] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1057.642374] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271961', 'volume_id': '3b174d32-3a39-4759-ba6c-6ea837f89b2e', 'name': 'volume-3b174d32-3a39-4759-ba6c-6ea837f89b2e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e44914cd-d8ae-4f43-9207-32cacf6d6553', 'attached_at': '', 'detached_at': '', 'volume_id': '3b174d32-3a39-4759-ba6c-6ea837f89b2e', 'serial': '3b174d32-3a39-4759-ba6c-6ea837f89b2e'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1057.643265] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9bc642-59c4-4f93-ad6c-a4fee34bf715 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.660178] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe6407ce-eae3-40c9-844c-e0a4fab1bd9c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.685400] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] volume-3b174d32-3a39-4759-ba6c-6ea837f89b2e/volume-3b174d32-3a39-4759-ba6c-6ea837f89b2e.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1057.685711] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1383b719-0c61-4ee0-a942-0323398b2649 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.703943] env[62952]: DEBUG oslo_vmware.api [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1057.703943] env[62952]: value = "task-1263779" [ 1057.703943] env[62952]: _type = "Task" [ 1057.703943] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.711897] env[62952]: DEBUG oslo_vmware.api [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263779, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.830975] env[62952]: DEBUG nova.network.neutron [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1057.890801] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 1a02fba4-aad0-4bd4-8183-a7f08d887a5a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1057.891164] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1057.891428] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1057.941111] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263778, 'name': Rename_Task, 'duration_secs': 0.363347} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.941413] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1057.941662] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-883a4efe-1cac-4e96-b1eb-2b3a5dd3b77a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.948048] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1057.948048] env[62952]: value = "task-1263780" [ 1057.948048] env[62952]: _type = "Task" [ 1057.948048] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.956406] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263780, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.017126] env[62952]: DEBUG nova.network.neutron [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Updating instance_info_cache with network_info: [{"id": "3ab932ab-2d08-4a0f-af17-e69bfb046602", "address": "fa:16:3e:08:4d:5f", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ab932ab-2d", "ovs_interfaceid": "3ab932ab-2d08-4a0f-af17-e69bfb046602", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.078021] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642a63e9-1cad-4b3d-bd26-79e52539fbc5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.079724] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "refresh_cache-e9e690b8-2fae-401d-aaad-0acef6e88ce3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.086714] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd222a9-b6ff-4862-a3d6-86c91946fdde {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.124819] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d17ee9-7a28-4c46-a8db-2b84304728f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.133270] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1786d536-ee86-4db5-b8c7-68b9be533d3d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.147527] env[62952]: DEBUG nova.compute.provider_tree [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.193373] env[62952]: DEBUG nova.network.neutron [req-486d9796-c15e-4246-88a7-701872487aac req-81d42f12-0ab4-482a-8f9d-e2085caa38ca service nova] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Updated VIF entry in instance network info cache for port d5013042-c6e4-4118-aaa6-3947b89a6fb3. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1058.193800] env[62952]: DEBUG nova.network.neutron [req-486d9796-c15e-4246-88a7-701872487aac req-81d42f12-0ab4-482a-8f9d-e2085caa38ca service nova] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Updating instance_info_cache with network_info: [{"id": "d5013042-c6e4-4118-aaa6-3947b89a6fb3", "address": "fa:16:3e:ac:f8:d7", "network": {"id": "bac9b3eb-e4d1-436c-b5d4-50db41886dfa", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1618686039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.237", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bff36f8c24b44ea8afc9094153087c5c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5013042-c6", "ovs_interfaceid": "d5013042-c6e4-4118-aaa6-3947b89a6fb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.213740] env[62952]: DEBUG oslo_vmware.api [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263779, 'name': ReconfigVM_Task, 'duration_secs': 0.417167} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.214073] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Reconfigured VM instance instance-00000055 to attach disk [datastore2] volume-3b174d32-3a39-4759-ba6c-6ea837f89b2e/volume-3b174d32-3a39-4759-ba6c-6ea837f89b2e.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1058.218906] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ffd9628-09ef-433e-94d8-7ddf867f3817 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.233417] env[62952]: DEBUG oslo_vmware.api [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1058.233417] env[62952]: value = "task-1263781" [ 1058.233417] env[62952]: _type = "Task" [ 1058.233417] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.240838] env[62952]: DEBUG oslo_vmware.api [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263781, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.458613] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263780, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.521137] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "refresh_cache-af261b32-457a-44a1-bab4-b62654b0be1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.521506] env[62952]: DEBUG nova.compute.manager [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Instance network_info: |[{"id": "3ab932ab-2d08-4a0f-af17-e69bfb046602", "address": "fa:16:3e:08:4d:5f", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ab932ab-2d", "ovs_interfaceid": "3ab932ab-2d08-4a0f-af17-e69bfb046602", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1058.521921] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:4d:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3ab932ab-2d08-4a0f-af17-e69bfb046602', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1058.529186] env[62952]: DEBUG oslo.service.loopingcall [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1058.529399] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1058.529645] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a657268f-d86c-4bd4-af06-42fa174f93bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.551016] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1058.551016] env[62952]: value = "task-1263782" [ 1058.551016] env[62952]: _type = "Task" [ 1058.551016] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.556697] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263782, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.598042] env[62952]: DEBUG nova.compute.manager [req-a5b136e5-13ff-4dd7-9f9f-a023b3161738 req-72a96fc7-0b51-46f6-bc07-07be643b1828 service nova] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Received event network-changed-3ab932ab-2d08-4a0f-af17-e69bfb046602 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1058.598264] env[62952]: DEBUG nova.compute.manager [req-a5b136e5-13ff-4dd7-9f9f-a023b3161738 req-72a96fc7-0b51-46f6-bc07-07be643b1828 service nova] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Refreshing instance network info cache due to event network-changed-3ab932ab-2d08-4a0f-af17-e69bfb046602. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1058.598398] env[62952]: DEBUG oslo_concurrency.lockutils [req-a5b136e5-13ff-4dd7-9f9f-a023b3161738 req-72a96fc7-0b51-46f6-bc07-07be643b1828 service nova] Acquiring lock "refresh_cache-af261b32-457a-44a1-bab4-b62654b0be1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.598591] env[62952]: DEBUG oslo_concurrency.lockutils [req-a5b136e5-13ff-4dd7-9f9f-a023b3161738 req-72a96fc7-0b51-46f6-bc07-07be643b1828 service nova] Acquired lock "refresh_cache-af261b32-457a-44a1-bab4-b62654b0be1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.598841] env[62952]: DEBUG nova.network.neutron [req-a5b136e5-13ff-4dd7-9f9f-a023b3161738 req-72a96fc7-0b51-46f6-bc07-07be643b1828 service nova] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Refreshing network info cache for port 3ab932ab-2d08-4a0f-af17-e69bfb046602 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1058.650618] env[62952]: DEBUG nova.scheduler.client.report [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1058.696966] env[62952]: DEBUG oslo_concurrency.lockutils [req-486d9796-c15e-4246-88a7-701872487aac req-81d42f12-0ab4-482a-8f9d-e2085caa38ca service nova] Releasing lock "refresh_cache-b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.744215] env[62952]: DEBUG oslo_vmware.api [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263781, 'name': ReconfigVM_Task, 'duration_secs': 0.138952} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.744643] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271961', 'volume_id': '3b174d32-3a39-4759-ba6c-6ea837f89b2e', 'name': 'volume-3b174d32-3a39-4759-ba6c-6ea837f89b2e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e44914cd-d8ae-4f43-9207-32cacf6d6553', 'attached_at': '', 'detached_at': '', 'volume_id': '3b174d32-3a39-4759-ba6c-6ea837f89b2e', 'serial': '3b174d32-3a39-4759-ba6c-6ea837f89b2e'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1058.960168] env[62952]: DEBUG oslo_vmware.api [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263780, 'name': PowerOnVM_Task, 'duration_secs': 0.869163} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.960623] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1058.960962] env[62952]: INFO nova.compute.manager [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Took 8.10 seconds to spawn the instance on the hypervisor. [ 1058.961321] env[62952]: DEBUG nova.compute.manager [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1058.962511] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee27c580-d404-4170-8931-6ce0bba047dd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.062397] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263782, 'name': CreateVM_Task, 'duration_secs': 0.425444} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.062879] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1059.063511] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.063927] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.064163] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1059.064492] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3afe51c-8c78-4fbd-b404-ad69c950a100 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.070453] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1059.070453] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5224189e-b086-00eb-5a73-098359952b33" [ 1059.070453] env[62952]: _type = "Task" [ 1059.070453] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.079706] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5224189e-b086-00eb-5a73-098359952b33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.155705] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1059.156074] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.812s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.156516] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.804s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.156859] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.159901] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.364s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.162807] env[62952]: INFO nova.compute.claims [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1059.189641] env[62952]: INFO nova.scheduler.client.report [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Deleted allocations for instance 05223053-90c3-433c-99b9-e252ef9a1935 [ 1059.310398] env[62952]: DEBUG nova.network.neutron [req-a5b136e5-13ff-4dd7-9f9f-a023b3161738 req-72a96fc7-0b51-46f6-bc07-07be643b1828 service nova] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Updated VIF entry in instance network info cache for port 3ab932ab-2d08-4a0f-af17-e69bfb046602. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1059.310797] env[62952]: DEBUG nova.network.neutron [req-a5b136e5-13ff-4dd7-9f9f-a023b3161738 req-72a96fc7-0b51-46f6-bc07-07be643b1828 service nova] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Updating instance_info_cache with network_info: [{"id": "3ab932ab-2d08-4a0f-af17-e69bfb046602", "address": "fa:16:3e:08:4d:5f", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ab932ab-2d", "ovs_interfaceid": "3ab932ab-2d08-4a0f-af17-e69bfb046602", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.479681] env[62952]: INFO nova.compute.manager [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Took 19.07 seconds to build instance. [ 1059.581394] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5224189e-b086-00eb-5a73-098359952b33, 'name': SearchDatastore_Task, 'duration_secs': 0.012955} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.581705] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.581944] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1059.582199] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.582349] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.582532] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1059.582799] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-87a40a97-fce7-4396-bcbe-d3cd3bd0604a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.590952] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1059.591229] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1059.594228] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4766d061-f6e6-4e30-b47f-ba04762f1213 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.598369] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102bc53c-7512-4d33-bc25-ca31d3279523 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.603904] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1059.603904] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521298b9-4117-5fda-8182-38450e514297" [ 1059.603904] env[62952]: _type = "Task" [ 1059.603904] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.617244] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Updating instance 'e9e690b8-2fae-401d-aaad-0acef6e88ce3' progress to 0 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1059.629075] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521298b9-4117-5fda-8182-38450e514297, 'name': SearchDatastore_Task, 'duration_secs': 0.01164} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.630119] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c29b530-0bd2-45a2-816b-cd455346ddfb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.635411] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1059.635411] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5219fb03-1bf5-053a-f9a4-4f2ed1043f4e" [ 1059.635411] env[62952]: _type = "Task" [ 1059.635411] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.642927] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5219fb03-1bf5-053a-f9a4-4f2ed1043f4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.700443] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2cc488f2-b8ec-4d6f-81d0-6658a0c2356a tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "05223053-90c3-433c-99b9-e252ef9a1935" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.792s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.782014] env[62952]: DEBUG nova.objects.instance [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lazy-loading 'flavor' on Instance uuid e44914cd-d8ae-4f43-9207-32cacf6d6553 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1059.812946] env[62952]: DEBUG oslo_concurrency.lockutils [req-a5b136e5-13ff-4dd7-9f9f-a023b3161738 req-72a96fc7-0b51-46f6-bc07-07be643b1828 service nova] Releasing lock "refresh_cache-af261b32-457a-44a1-bab4-b62654b0be1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.980424] env[62952]: DEBUG oslo_concurrency.lockutils [None req-70114494-6eeb-4e0a-9005-e246861d8f2c tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "1a38bdcd-5537-46f5-94f6-a8e358e13121" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.583s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.128404] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1060.129095] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cf8a7daf-f054-4e56-b27b-160208321764 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.137884] env[62952]: DEBUG oslo_vmware.api [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1060.137884] env[62952]: value = "task-1263783" [ 1060.137884] env[62952]: _type = "Task" [ 1060.137884] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.156190] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5219fb03-1bf5-053a-f9a4-4f2ed1043f4e, 'name': SearchDatastore_Task, 'duration_secs': 0.008827} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.161155] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.161547] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] af261b32-457a-44a1-bab4-b62654b0be1f/af261b32-457a-44a1-bab4-b62654b0be1f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1060.161968] env[62952]: DEBUG oslo_vmware.api [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263783, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.162300] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39b2c434-ebf3-4a88-ad31-86890b4d7335 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.174037] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1060.174037] env[62952]: value = "task-1263784" [ 1060.174037] env[62952]: _type = "Task" [ 1060.174037] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.187216] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263784, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.288677] env[62952]: DEBUG oslo_concurrency.lockutils [None req-40de6dfd-1447-44ba-916d-b18c03b19c89 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.253s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.400436] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cccf42ad-2946-431f-93cc-e439d89eaf00 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.416024] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f3ef73-0d70-4057-b60d-bb217f22773a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.458895] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c842fc71-901c-4bb4-aa4d-043fbfd1447b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.467701] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b29cfa3-69cb-441c-b314-2cd41228cffd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.483956] env[62952]: DEBUG nova.compute.provider_tree [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.655960] env[62952]: DEBUG oslo_vmware.api [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263783, 'name': PowerOffVM_Task, 'duration_secs': 0.198563} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.656481] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1060.656823] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Updating instance 'e9e690b8-2fae-401d-aaad-0acef6e88ce3' progress to 17 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1060.691086] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263784, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500423} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.691822] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] af261b32-457a-44a1-bab4-b62654b0be1f/af261b32-457a-44a1-bab4-b62654b0be1f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1060.695421] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1060.695421] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf120230-ec6c-46bc-9d22-822191d391fe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.703360] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1060.703360] env[62952]: value = "task-1263785" [ 1060.703360] env[62952]: _type = "Task" [ 1060.703360] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.709567] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263785, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.988060] env[62952]: DEBUG nova.scheduler.client.report [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1061.155512] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.155777] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.163022] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1061.163495] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1061.163730] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1061.163939] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1061.164124] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1061.164262] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1061.164471] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1061.164639] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1061.164808] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1061.164975] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1061.165174] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1061.171312] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-899ba48e-2fae-4f07-9755-7fdccce8d784 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.189021] env[62952]: DEBUG oslo_vmware.api [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1061.189021] env[62952]: value = "task-1263786" [ 1061.189021] env[62952]: _type = "Task" [ 1061.189021] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.198403] env[62952]: DEBUG oslo_vmware.api [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263786, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.211065] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263785, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078561} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.211357] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1061.212214] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9aaa84-d0a7-4ae5-b0b0-25473e2bd778 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.235040] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] af261b32-457a-44a1-bab4-b62654b0be1f/af261b32-457a-44a1-bab4-b62654b0be1f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1061.235631] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-589b59a8-3a22-4efc-aed8-e3950dee411e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.257617] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1061.257617] env[62952]: value = "task-1263787" [ 1061.257617] env[62952]: _type = "Task" [ 1061.257617] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.267782] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263787, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.321094] env[62952]: DEBUG oslo_concurrency.lockutils [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "1a38bdcd-5537-46f5-94f6-a8e358e13121" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.321389] env[62952]: DEBUG oslo_concurrency.lockutils [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "1a38bdcd-5537-46f5-94f6-a8e358e13121" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.321621] env[62952]: DEBUG oslo_concurrency.lockutils [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "1a38bdcd-5537-46f5-94f6-a8e358e13121-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.321810] env[62952]: DEBUG oslo_concurrency.lockutils [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "1a38bdcd-5537-46f5-94f6-a8e358e13121-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.321984] env[62952]: DEBUG oslo_concurrency.lockutils [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "1a38bdcd-5537-46f5-94f6-a8e358e13121-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.324551] env[62952]: INFO nova.compute.manager [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Terminating instance [ 1061.326896] env[62952]: DEBUG nova.compute.manager [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1061.327124] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1061.327947] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789ddf12-9f6a-4770-906c-0f8e09b2b24e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.337500] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1061.337766] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bbf16ab2-ada5-4828-b067-0f3a698852ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.344380] env[62952]: DEBUG oslo_vmware.api [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1061.344380] env[62952]: value = "task-1263788" [ 1061.344380] env[62952]: _type = "Task" [ 1061.344380] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.351954] env[62952]: DEBUG oslo_vmware.api [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263788, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.493430] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.494364] env[62952]: DEBUG nova.compute.manager [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1061.496886] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.826s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.497673] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.518122] env[62952]: INFO nova.scheduler.client.report [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Deleted allocations for instance 7ba1587d-edaf-4dce-a224-6b2ac22cfecd [ 1061.660063] env[62952]: DEBUG nova.compute.utils [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1061.699357] env[62952]: DEBUG oslo_vmware.api [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263786, 'name': ReconfigVM_Task, 'duration_secs': 0.322382} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.699677] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Updating instance 'e9e690b8-2fae-401d-aaad-0acef6e88ce3' progress to 33 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1061.767279] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263787, 'name': ReconfigVM_Task, 'duration_secs': 0.410842} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.767628] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Reconfigured VM instance instance-00000061 to attach disk [datastore2] af261b32-457a-44a1-bab4-b62654b0be1f/af261b32-457a-44a1-bab4-b62654b0be1f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1061.768237] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e2942130-4008-43e6-ae84-e53eb7be330a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.774622] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1061.774622] env[62952]: value = "task-1263789" [ 1061.774622] env[62952]: _type = "Task" [ 1061.774622] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.781975] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263789, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.854284] env[62952]: DEBUG oslo_vmware.api [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263788, 'name': PowerOffVM_Task, 'duration_secs': 0.248519} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.854562] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1061.854776] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1061.855044] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-56fc0817-4d3d-49b0-801e-ea18c061de7a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.918109] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1061.918336] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1061.918609] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleting the datastore file [datastore2] 1a38bdcd-5537-46f5-94f6-a8e358e13121 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1061.918887] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0cbec380-24f2-45d3-8040-7035c94df495 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.927048] env[62952]: DEBUG oslo_vmware.api [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1061.927048] env[62952]: value = "task-1263791" [ 1061.927048] env[62952]: _type = "Task" [ 1061.927048] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.934819] env[62952]: DEBUG oslo_vmware.api [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263791, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.992675] env[62952]: DEBUG nova.compute.manager [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Stashing vm_state: active {{(pid=62952) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1062.001943] env[62952]: DEBUG nova.compute.utils [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1062.003983] env[62952]: DEBUG nova.compute.manager [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1062.003983] env[62952]: DEBUG nova.network.neutron [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1062.027015] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1a1a07ed-024c-4451-bb9c-59ed99905ed6 tempest-ServerRescueTestJSON-1891236444 tempest-ServerRescueTestJSON-1891236444-project-member] Lock "7ba1587d-edaf-4dce-a224-6b2ac22cfecd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.884s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.070911] env[62952]: DEBUG nova.policy [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0f483bd2140d43399500808fd217449c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '730bb2e1e971424c982bdec105343de5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1062.163393] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.205900] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1062.206186] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1062.206385] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1062.206620] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1062.206775] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1062.206930] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1062.207152] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1062.207316] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1062.207489] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1062.207697] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1062.208114] env[62952]: DEBUG nova.virt.hardware [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1062.213780] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Reconfiguring VM instance instance-0000005d to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1062.214094] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f2f0b2c-290e-4060-bc56-0e0fe8d63e4f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.233736] env[62952]: DEBUG oslo_vmware.api [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1062.233736] env[62952]: value = "task-1263792" [ 1062.233736] env[62952]: _type = "Task" [ 1062.233736] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.241891] env[62952]: DEBUG oslo_vmware.api [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263792, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.286354] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263789, 'name': Rename_Task, 'duration_secs': 0.151909} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.286697] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1062.287083] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0fdc148-4c36-4508-a321-79c9697a94b9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.293644] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1062.293644] env[62952]: value = "task-1263793" [ 1062.293644] env[62952]: _type = "Task" [ 1062.293644] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.302828] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263793, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.436990] env[62952]: DEBUG oslo_vmware.api [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263791, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172383} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.437277] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1062.437467] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1062.437688] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1062.437887] env[62952]: INFO nova.compute.manager [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1062.438146] env[62952]: DEBUG oslo.service.loopingcall [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1062.438339] env[62952]: DEBUG nova.compute.manager [-] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1062.438461] env[62952]: DEBUG nova.network.neutron [-] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1062.492323] env[62952]: DEBUG nova.network.neutron [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Successfully created port: 87c0fa2c-a6cf-4539-a9d4-3c4394c10068 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1062.506578] env[62952]: DEBUG nova.compute.manager [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1062.514824] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.515152] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.721940] env[62952]: DEBUG nova.compute.manager [req-c40d67ce-a6f5-44a9-8beb-de1d39a1e8dd req-e0f8057c-cf8a-43aa-a17e-dfcffc7bc61b service nova] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Received event network-vif-deleted-70b1d946-ae76-4aba-9c14-ecc65d044630 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1062.722186] env[62952]: INFO nova.compute.manager [req-c40d67ce-a6f5-44a9-8beb-de1d39a1e8dd req-e0f8057c-cf8a-43aa-a17e-dfcffc7bc61b service nova] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Neutron deleted interface 70b1d946-ae76-4aba-9c14-ecc65d044630; detaching it from the instance and deleting it from the info cache [ 1062.722376] env[62952]: DEBUG nova.network.neutron [req-c40d67ce-a6f5-44a9-8beb-de1d39a1e8dd req-e0f8057c-cf8a-43aa-a17e-dfcffc7bc61b service nova] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.747034] env[62952]: DEBUG oslo_vmware.api [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263792, 'name': ReconfigVM_Task, 'duration_secs': 0.188744} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.747433] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Reconfigured VM instance instance-0000005d to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1062.748274] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d621c638-0971-4548-a816-f7e7b368e66a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.772730] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] e9e690b8-2fae-401d-aaad-0acef6e88ce3/e9e690b8-2fae-401d-aaad-0acef6e88ce3.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1062.773389] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b1f58b47-16af-431c-a7b7-6ec3ce0f2a16 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.793014] env[62952]: DEBUG oslo_vmware.api [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1062.793014] env[62952]: value = "task-1263794" [ 1062.793014] env[62952]: _type = "Task" [ 1062.793014] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.804422] env[62952]: DEBUG oslo_vmware.api [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263794, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.807488] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263793, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.021546] env[62952]: INFO nova.compute.claims [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1063.200295] env[62952]: DEBUG nova.network.neutron [-] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.225341] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8cbed6e1-1060-46e4-8b82-7e1add05944f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.235429] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e643ed9-9e59-40e3-9f4d-045941bfe5ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.245978] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.246403] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.246646] env[62952]: INFO nova.compute.manager [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Attaching volume 1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e to /dev/sdc [ 1063.267572] env[62952]: DEBUG nova.compute.manager [req-c40d67ce-a6f5-44a9-8beb-de1d39a1e8dd req-e0f8057c-cf8a-43aa-a17e-dfcffc7bc61b service nova] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Detach interface failed, port_id=70b1d946-ae76-4aba-9c14-ecc65d044630, reason: Instance 1a38bdcd-5537-46f5-94f6-a8e358e13121 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1063.282433] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0da9a82-c3bf-41cb-a48a-5e95dbfc27aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.289280] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e78941ad-4a1c-41e2-b3d1-4589cf84e1c8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.302794] env[62952]: DEBUG nova.virt.block_device [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Updating existing volume attachment record: b2a8648f-0525-44c1-805d-631760940bdc {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1063.307918] env[62952]: DEBUG oslo_vmware.api [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263794, 'name': ReconfigVM_Task, 'duration_secs': 0.294154} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.308196] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Reconfigured VM instance instance-0000005d to attach disk [datastore1] e9e690b8-2fae-401d-aaad-0acef6e88ce3/e9e690b8-2fae-401d-aaad-0acef6e88ce3.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1063.308761] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Updating instance 'e9e690b8-2fae-401d-aaad-0acef6e88ce3' progress to 50 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1063.314307] env[62952]: DEBUG oslo_vmware.api [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263793, 'name': PowerOnVM_Task, 'duration_secs': 0.51718} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.314858] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1063.315113] env[62952]: INFO nova.compute.manager [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Took 7.28 seconds to spawn the instance on the hypervisor. [ 1063.315303] env[62952]: DEBUG nova.compute.manager [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1063.316049] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc930b58-32ee-4a50-bd3f-4e89471e40a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.517334] env[62952]: DEBUG nova.compute.manager [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1063.529723] env[62952]: INFO nova.compute.resource_tracker [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating resource usage from migration 8ba8486b-d3a9-43a5-9640-d6bc578ae322 [ 1063.546388] env[62952]: DEBUG nova.virt.hardware [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1063.546388] env[62952]: DEBUG nova.virt.hardware [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1063.546388] env[62952]: DEBUG nova.virt.hardware [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1063.546634] env[62952]: DEBUG nova.virt.hardware [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1063.546680] env[62952]: DEBUG nova.virt.hardware [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1063.546835] env[62952]: DEBUG nova.virt.hardware [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1063.547059] env[62952]: DEBUG nova.virt.hardware [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1063.547224] env[62952]: DEBUG nova.virt.hardware [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1063.547389] env[62952]: DEBUG nova.virt.hardware [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1063.547581] env[62952]: DEBUG nova.virt.hardware [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1063.548010] env[62952]: DEBUG nova.virt.hardware [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1063.548747] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e6259d-6e26-44c0-80d9-066bf13367a3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.559681] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee76c9d3-e163-465e-b570-869e715362c8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.704046] env[62952]: INFO nova.compute.manager [-] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Took 1.27 seconds to deallocate network for instance. [ 1063.727037] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c37a60-4e27-433d-8cb9-67633c895fb8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.735464] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b40b9df-2938-4833-be9f-4253456582df {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.773790] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3799c5b2-d0b9-48a4-9b29-07bd6642cfc2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.779865] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b009d727-a25d-4348-a54b-efb4bfdcea79 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.794728] env[62952]: DEBUG nova.compute.provider_tree [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1063.820334] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed0ead2-4890-4b03-ba4c-35992b75ad36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.847739] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a01834d-4abb-4d6f-8daf-2f2b902faf97 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.851545] env[62952]: INFO nova.compute.manager [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Took 18.64 seconds to build instance. [ 1063.874416] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Updating instance 'e9e690b8-2fae-401d-aaad-0acef6e88ce3' progress to 67 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1064.211219] env[62952]: DEBUG oslo_concurrency.lockutils [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.258377] env[62952]: DEBUG nova.compute.manager [req-31f97254-b571-45e9-b5bd-fa2b3c6c8297 req-01e58c80-ff0a-47a7-b89d-4f3231545b0f service nova] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Received event network-vif-plugged-87c0fa2c-a6cf-4539-a9d4-3c4394c10068 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1064.258707] env[62952]: DEBUG oslo_concurrency.lockutils [req-31f97254-b571-45e9-b5bd-fa2b3c6c8297 req-01e58c80-ff0a-47a7-b89d-4f3231545b0f service nova] Acquiring lock "1a02fba4-aad0-4bd4-8183-a7f08d887a5a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.258973] env[62952]: DEBUG oslo_concurrency.lockutils [req-31f97254-b571-45e9-b5bd-fa2b3c6c8297 req-01e58c80-ff0a-47a7-b89d-4f3231545b0f service nova] Lock "1a02fba4-aad0-4bd4-8183-a7f08d887a5a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.259179] env[62952]: DEBUG oslo_concurrency.lockutils [req-31f97254-b571-45e9-b5bd-fa2b3c6c8297 req-01e58c80-ff0a-47a7-b89d-4f3231545b0f service nova] Lock "1a02fba4-aad0-4bd4-8183-a7f08d887a5a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.259341] env[62952]: DEBUG nova.compute.manager [req-31f97254-b571-45e9-b5bd-fa2b3c6c8297 req-01e58c80-ff0a-47a7-b89d-4f3231545b0f service nova] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] No waiting events found dispatching network-vif-plugged-87c0fa2c-a6cf-4539-a9d4-3c4394c10068 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1064.259519] env[62952]: WARNING nova.compute.manager [req-31f97254-b571-45e9-b5bd-fa2b3c6c8297 req-01e58c80-ff0a-47a7-b89d-4f3231545b0f service nova] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Received unexpected event network-vif-plugged-87c0fa2c-a6cf-4539-a9d4-3c4394c10068 for instance with vm_state building and task_state spawning. [ 1064.298981] env[62952]: DEBUG nova.scheduler.client.report [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1064.348089] env[62952]: DEBUG nova.network.neutron [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Successfully updated port: 87c0fa2c-a6cf-4539-a9d4-3c4394c10068 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1064.355332] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f4f9fa7-5b6d-4245-81b0-ade86886882b tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "af261b32-457a-44a1-bab4-b62654b0be1f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.154s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.402078] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ba441c3-1d0e-4974-8fa4-11a52b3a878a tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "af261b32-457a-44a1-bab4-b62654b0be1f" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.402348] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ba441c3-1d0e-4974-8fa4-11a52b3a878a tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "af261b32-457a-44a1-bab4-b62654b0be1f" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.402517] env[62952]: DEBUG nova.compute.manager [None req-9ba441c3-1d0e-4974-8fa4-11a52b3a878a tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1064.403435] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e23f7c4-8525-4cd0-bc56-4d0c407a03d0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.410192] env[62952]: DEBUG nova.compute.manager [None req-9ba441c3-1d0e-4974-8fa4-11a52b3a878a tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62952) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1064.410751] env[62952]: DEBUG nova.objects.instance [None req-9ba441c3-1d0e-4974-8fa4-11a52b3a878a tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lazy-loading 'flavor' on Instance uuid af261b32-457a-44a1-bab4-b62654b0be1f {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1064.415879] env[62952]: DEBUG nova.network.neutron [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Port cff4c81c-513a-42a4-b6f5-94b4367fa0ea binding to destination host cpu-1 is already ACTIVE {{(pid=62952) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1064.803783] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.289s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.804100] env[62952]: INFO nova.compute.manager [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Migrating [ 1064.810748] env[62952]: DEBUG oslo_concurrency.lockutils [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.600s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.810940] env[62952]: DEBUG nova.objects.instance [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lazy-loading 'resources' on Instance uuid 1a38bdcd-5537-46f5-94f6-a8e358e13121 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1064.851294] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Acquiring lock "refresh_cache-1a02fba4-aad0-4bd4-8183-a7f08d887a5a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.851455] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Acquired lock "refresh_cache-1a02fba4-aad0-4bd4-8183-a7f08d887a5a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.851627] env[62952]: DEBUG nova.network.neutron [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1064.916033] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ba441c3-1d0e-4974-8fa4-11a52b3a878a tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1064.916339] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-84005bf7-252d-4268-9108-75bfd44033e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.927994] env[62952]: DEBUG oslo_vmware.api [None req-9ba441c3-1d0e-4974-8fa4-11a52b3a878a tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1064.927994] env[62952]: value = "task-1263796" [ 1064.927994] env[62952]: _type = "Task" [ 1064.927994] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.935924] env[62952]: DEBUG oslo_vmware.api [None req-9ba441c3-1d0e-4974-8fa4-11a52b3a878a tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263796, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.318874] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.319097] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.319690] env[62952]: DEBUG nova.network.neutron [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1065.385711] env[62952]: DEBUG nova.network.neutron [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1065.446840] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.446840] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.446840] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.454188] env[62952]: DEBUG oslo_vmware.api [None req-9ba441c3-1d0e-4974-8fa4-11a52b3a878a tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263796, 'name': PowerOffVM_Task, 'duration_secs': 0.178604} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.454689] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ba441c3-1d0e-4974-8fa4-11a52b3a878a tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1065.454906] env[62952]: DEBUG nova.compute.manager [None req-9ba441c3-1d0e-4974-8fa4-11a52b3a878a tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.455732] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ef06da-5424-46e6-8f33-aadfedfb1c3f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.513033] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49b30fc3-2767-4e6a-a59d-325df7cbcd0d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.523326] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c15cda-aa52-47f7-b9b6-c59f8d79d031 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.553297] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2537ef27-ed73-4987-856c-a1c67d2386bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.556757] env[62952]: DEBUG nova.network.neutron [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Updating instance_info_cache with network_info: [{"id": "87c0fa2c-a6cf-4539-a9d4-3c4394c10068", "address": "fa:16:3e:30:fe:8a", "network": {"id": "0228ecc2-4820-4752-8f85-9478b687180e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-504638304-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "730bb2e1e971424c982bdec105343de5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed8a78a1-87dc-488e-a092-afd1c2a2ddde", "external-id": "nsx-vlan-transportzone-21", "segmentation_id": 21, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87c0fa2c-a6", "ovs_interfaceid": "87c0fa2c-a6cf-4539-a9d4-3c4394c10068", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.562995] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca16357e-8b75-4a8e-842c-ff6a9a9b538e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.576737] env[62952]: DEBUG nova.compute.provider_tree [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.972589] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9ba441c3-1d0e-4974-8fa4-11a52b3a878a tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "af261b32-457a-44a1-bab4-b62654b0be1f" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.570s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.024014] env[62952]: DEBUG nova.network.neutron [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance_info_cache with network_info: [{"id": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "address": "fa:16:3e:eb:c2:8d", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e2e9d1e-dc", "ovs_interfaceid": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.059137] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Releasing lock "refresh_cache-1a02fba4-aad0-4bd4-8183-a7f08d887a5a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.059473] env[62952]: DEBUG nova.compute.manager [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Instance network_info: |[{"id": "87c0fa2c-a6cf-4539-a9d4-3c4394c10068", "address": "fa:16:3e:30:fe:8a", "network": {"id": "0228ecc2-4820-4752-8f85-9478b687180e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-504638304-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "730bb2e1e971424c982bdec105343de5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed8a78a1-87dc-488e-a092-afd1c2a2ddde", "external-id": "nsx-vlan-transportzone-21", "segmentation_id": 21, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87c0fa2c-a6", "ovs_interfaceid": "87c0fa2c-a6cf-4539-a9d4-3c4394c10068", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1066.059865] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:fe:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ed8a78a1-87dc-488e-a092-afd1c2a2ddde', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '87c0fa2c-a6cf-4539-a9d4-3c4394c10068', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1066.067128] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Creating folder: Project (730bb2e1e971424c982bdec105343de5). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1066.067682] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dd014eb9-38ba-4e53-9925-af4e2facd72a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.079040] env[62952]: DEBUG nova.scheduler.client.report [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1066.083053] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Created folder: Project (730bb2e1e971424c982bdec105343de5) in parent group-v271811. [ 1066.083053] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Creating folder: Instances. Parent ref: group-v271964. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1066.083172] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2e68bd0b-5296-4e93-a94e-85d5a76d28c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.091954] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Created folder: Instances in parent group-v271964. [ 1066.092190] env[62952]: DEBUG oslo.service.loopingcall [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.092837] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1066.093049] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fbde0db3-1902-4290-ab40-f586d8b55581 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.112290] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1066.112290] env[62952]: value = "task-1263800" [ 1066.112290] env[62952]: _type = "Task" [ 1066.112290] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.119579] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263800, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.287656] env[62952]: DEBUG nova.compute.manager [req-c88d612a-5903-41ac-b65b-a97c8aef5665 req-8094c9c3-98a3-4345-8d0d-d0cbb1dfb4c8 service nova] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Received event network-changed-87c0fa2c-a6cf-4539-a9d4-3c4394c10068 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1066.287656] env[62952]: DEBUG nova.compute.manager [req-c88d612a-5903-41ac-b65b-a97c8aef5665 req-8094c9c3-98a3-4345-8d0d-d0cbb1dfb4c8 service nova] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Refreshing instance network info cache due to event network-changed-87c0fa2c-a6cf-4539-a9d4-3c4394c10068. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1066.287656] env[62952]: DEBUG oslo_concurrency.lockutils [req-c88d612a-5903-41ac-b65b-a97c8aef5665 req-8094c9c3-98a3-4345-8d0d-d0cbb1dfb4c8 service nova] Acquiring lock "refresh_cache-1a02fba4-aad0-4bd4-8183-a7f08d887a5a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.287656] env[62952]: DEBUG oslo_concurrency.lockutils [req-c88d612a-5903-41ac-b65b-a97c8aef5665 req-8094c9c3-98a3-4345-8d0d-d0cbb1dfb4c8 service nova] Acquired lock "refresh_cache-1a02fba4-aad0-4bd4-8183-a7f08d887a5a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.287656] env[62952]: DEBUG nova.network.neutron [req-c88d612a-5903-41ac-b65b-a97c8aef5665 req-8094c9c3-98a3-4345-8d0d-d0cbb1dfb4c8 service nova] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Refreshing network info cache for port 87c0fa2c-a6cf-4539-a9d4-3c4394c10068 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1066.481388] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "refresh_cache-e9e690b8-2fae-401d-aaad-0acef6e88ce3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.481388] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "refresh_cache-e9e690b8-2fae-401d-aaad-0acef6e88ce3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.481388] env[62952]: DEBUG nova.network.neutron [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1066.526519] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.585573] env[62952]: DEBUG oslo_concurrency.lockutils [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.774s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.601369] env[62952]: INFO nova.scheduler.client.report [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleted allocations for instance 1a38bdcd-5537-46f5-94f6-a8e358e13121 [ 1066.622053] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263800, 'name': CreateVM_Task, 'duration_secs': 0.306757} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.622270] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1066.622963] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.623190] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.623523] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1066.623781] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-257ab983-d9c9-47ed-b401-8f98b064e1ba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.628538] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Waiting for the task: (returnval){ [ 1066.628538] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52239cce-b0df-984c-14b1-3051bd434cf2" [ 1066.628538] env[62952]: _type = "Task" [ 1066.628538] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.636217] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52239cce-b0df-984c-14b1-3051bd434cf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.829681] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "af261b32-457a-44a1-bab4-b62654b0be1f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.829938] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "af261b32-457a-44a1-bab4-b62654b0be1f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.830170] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "af261b32-457a-44a1-bab4-b62654b0be1f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.830361] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "af261b32-457a-44a1-bab4-b62654b0be1f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.830534] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "af261b32-457a-44a1-bab4-b62654b0be1f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.832980] env[62952]: INFO nova.compute.manager [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Terminating instance [ 1066.834633] env[62952]: DEBUG nova.compute.manager [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1066.834830] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1066.835676] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b8169f-af89-4d98-bed7-ea72975da828 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.842784] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1066.843021] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f107a98b-5801-4836-b3fc-45ecf43bcd77 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.903548] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1066.903773] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1066.903963] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleting the datastore file [datastore2] af261b32-457a-44a1-bab4-b62654b0be1f {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1066.904281] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-105d4c89-08e7-4f60-936f-b9e52eb8002a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.911018] env[62952]: DEBUG oslo_vmware.api [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1066.911018] env[62952]: value = "task-1263802" [ 1066.911018] env[62952]: _type = "Task" [ 1066.911018] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.918823] env[62952]: DEBUG oslo_vmware.api [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263802, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.953846] env[62952]: DEBUG oslo_concurrency.lockutils [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "1941365e-d562-4d99-89eb-226a07e52071" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.954095] env[62952]: DEBUG oslo_concurrency.lockutils [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "1941365e-d562-4d99-89eb-226a07e52071" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.992286] env[62952]: DEBUG nova.network.neutron [req-c88d612a-5903-41ac-b65b-a97c8aef5665 req-8094c9c3-98a3-4345-8d0d-d0cbb1dfb4c8 service nova] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Updated VIF entry in instance network info cache for port 87c0fa2c-a6cf-4539-a9d4-3c4394c10068. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1066.992708] env[62952]: DEBUG nova.network.neutron [req-c88d612a-5903-41ac-b65b-a97c8aef5665 req-8094c9c3-98a3-4345-8d0d-d0cbb1dfb4c8 service nova] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Updating instance_info_cache with network_info: [{"id": "87c0fa2c-a6cf-4539-a9d4-3c4394c10068", "address": "fa:16:3e:30:fe:8a", "network": {"id": "0228ecc2-4820-4752-8f85-9478b687180e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-504638304-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "730bb2e1e971424c982bdec105343de5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed8a78a1-87dc-488e-a092-afd1c2a2ddde", "external-id": "nsx-vlan-transportzone-21", "segmentation_id": 21, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87c0fa2c-a6", "ovs_interfaceid": "87c0fa2c-a6cf-4539-a9d4-3c4394c10068", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.108712] env[62952]: DEBUG oslo_concurrency.lockutils [None req-abd404f9-b148-432f-8363-ec1038069bb2 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "1a38bdcd-5537-46f5-94f6-a8e358e13121" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.787s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.138971] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52239cce-b0df-984c-14b1-3051bd434cf2, 'name': SearchDatastore_Task, 'duration_secs': 0.009306} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.139304] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.139546] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1067.139780] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.139930] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.140132] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1067.140390] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fbf1a725-d6d9-4a34-82b7-fb689157f2fa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.147976] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1067.148168] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1067.148871] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c56158a-29dd-4672-b4a5-2a491574ac69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.153772] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Waiting for the task: (returnval){ [ 1067.153772] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52679063-d3eb-65b0-7f82-f204eb1bf221" [ 1067.153772] env[62952]: _type = "Task" [ 1067.153772] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.160965] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52679063-d3eb-65b0-7f82-f204eb1bf221, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.183204] env[62952]: DEBUG nova.network.neutron [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Updating instance_info_cache with network_info: [{"id": "cff4c81c-513a-42a4-b6f5-94b4367fa0ea", "address": "fa:16:3e:3a:e4:22", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcff4c81c-51", "ovs_interfaceid": "cff4c81c-513a-42a4-b6f5-94b4367fa0ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.420907] env[62952]: DEBUG oslo_vmware.api [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263802, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121856} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.421240] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1067.421446] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1067.421629] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1067.421805] env[62952]: INFO nova.compute.manager [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1067.422113] env[62952]: DEBUG oslo.service.loopingcall [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1067.422439] env[62952]: DEBUG nova.compute.manager [-] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1067.422439] env[62952]: DEBUG nova.network.neutron [-] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1067.459940] env[62952]: DEBUG nova.compute.utils [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1067.497985] env[62952]: DEBUG oslo_concurrency.lockutils [req-c88d612a-5903-41ac-b65b-a97c8aef5665 req-8094c9c3-98a3-4345-8d0d-d0cbb1dfb4c8 service nova] Releasing lock "refresh_cache-1a02fba4-aad0-4bd4-8183-a7f08d887a5a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.665325] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52679063-d3eb-65b0-7f82-f204eb1bf221, 'name': SearchDatastore_Task, 'duration_secs': 0.007754} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.665653] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bb5dcc8-2499-467d-8d96-73f33081907a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.670811] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Waiting for the task: (returnval){ [ 1067.670811] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5279b750-f121-e98d-e576-514c00dc674a" [ 1067.670811] env[62952]: _type = "Task" [ 1067.670811] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.679051] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5279b750-f121-e98d-e576-514c00dc674a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.685636] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "refresh_cache-e9e690b8-2fae-401d-aaad-0acef6e88ce3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.854039] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1067.854343] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271963', 'volume_id': '1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e', 'name': 'volume-1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e44914cd-d8ae-4f43-9207-32cacf6d6553', 'attached_at': '', 'detached_at': '', 'volume_id': '1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e', 'serial': '1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1067.855246] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a08ad22-3830-4682-91a0-f888a253386c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.871273] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f413640b-3c53-4144-94b9-3b4ef87b6e1f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.902169] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] volume-1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e/volume-1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1067.902169] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f57ed406-2f87-4421-b3b4-f33010e66672 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.922284] env[62952]: DEBUG oslo_vmware.api [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1067.922284] env[62952]: value = "task-1263803" [ 1067.922284] env[62952]: _type = "Task" [ 1067.922284] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.932935] env[62952]: DEBUG oslo_vmware.api [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263803, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.963760] env[62952]: DEBUG oslo_concurrency.lockutils [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "1941365e-d562-4d99-89eb-226a07e52071" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.042157] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184c3b33-18d9-4b3a-9ef0-ce9f5587276d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.066283] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance '6e289c20-8dd2-4680-b816-9bbe82fe7aec' progress to 0 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1068.137773] env[62952]: DEBUG nova.compute.manager [req-697bb0cc-9c78-4079-9c83-3c0badeb28e4 req-069bd11f-d241-4797-83bc-db9dd6bf935f service nova] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Received event network-vif-deleted-3ab932ab-2d08-4a0f-af17-e69bfb046602 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1068.139750] env[62952]: INFO nova.compute.manager [req-697bb0cc-9c78-4079-9c83-3c0badeb28e4 req-069bd11f-d241-4797-83bc-db9dd6bf935f service nova] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Neutron deleted interface 3ab932ab-2d08-4a0f-af17-e69bfb046602; detaching it from the instance and deleting it from the info cache [ 1068.139750] env[62952]: DEBUG nova.network.neutron [req-697bb0cc-9c78-4079-9c83-3c0badeb28e4 req-069bd11f-d241-4797-83bc-db9dd6bf935f service nova] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.183330] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5279b750-f121-e98d-e576-514c00dc674a, 'name': SearchDatastore_Task, 'duration_secs': 0.013814} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.183501] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.183770] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 1a02fba4-aad0-4bd4-8183-a7f08d887a5a/1a02fba4-aad0-4bd4-8183-a7f08d887a5a.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1068.184052] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fa09f08c-3621-43cc-ba17-6f1f53fce921 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.193931] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Waiting for the task: (returnval){ [ 1068.193931] env[62952]: value = "task-1263804" [ 1068.193931] env[62952]: _type = "Task" [ 1068.193931] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.205018] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': task-1263804, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.211761] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5216d9ff-04c0-4a7f-a891-68614eb8dceb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.229803] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-454571f9-42aa-4299-b21f-7597c34706a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.232663] env[62952]: DEBUG nova.network.neutron [-] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.240783] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Updating instance 'e9e690b8-2fae-401d-aaad-0acef6e88ce3' progress to 83 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1068.434345] env[62952]: DEBUG oslo_vmware.api [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263803, 'name': ReconfigVM_Task, 'duration_secs': 0.356751} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.434648] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Reconfigured VM instance instance-00000055 to attach disk [datastore1] volume-1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e/volume-1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1068.439718] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82e3e82a-983a-4251-8b77-540cdb0d1afd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.456085] env[62952]: DEBUG oslo_vmware.api [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1068.456085] env[62952]: value = "task-1263805" [ 1068.456085] env[62952]: _type = "Task" [ 1068.456085] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.465973] env[62952]: DEBUG oslo_vmware.api [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263805, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.525571] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "f28d0043-236e-4df4-8478-e4b1befb6a7d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.525841] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "f28d0043-236e-4df4-8478-e4b1befb6a7d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.573018] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1068.573822] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb090b8e-af44-43f9-be28-7203df22bd05 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.581282] env[62952]: DEBUG oslo_vmware.api [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1068.581282] env[62952]: value = "task-1263806" [ 1068.581282] env[62952]: _type = "Task" [ 1068.581282] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.591125] env[62952]: DEBUG oslo_vmware.api [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263806, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.642410] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fcd557ba-f4ef-4d7e-9bcd-d726c2a30eda {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.651173] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f057fc7-00d5-410a-8627-30f42a3e31db {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.681297] env[62952]: DEBUG nova.compute.manager [req-697bb0cc-9c78-4079-9c83-3c0badeb28e4 req-069bd11f-d241-4797-83bc-db9dd6bf935f service nova] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Detach interface failed, port_id=3ab932ab-2d08-4a0f-af17-e69bfb046602, reason: Instance af261b32-457a-44a1-bab4-b62654b0be1f could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1068.705942] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': task-1263804, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.433845} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.706232] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 1a02fba4-aad0-4bd4-8183-a7f08d887a5a/1a02fba4-aad0-4bd4-8183-a7f08d887a5a.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1068.706411] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1068.706674] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2fab40c6-92d9-49e4-b7c7-37018afbabfc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.714162] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Waiting for the task: (returnval){ [ 1068.714162] env[62952]: value = "task-1263807" [ 1068.714162] env[62952]: _type = "Task" [ 1068.714162] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.722788] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': task-1263807, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.736123] env[62952]: INFO nova.compute.manager [-] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Took 1.31 seconds to deallocate network for instance. [ 1068.745076] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1068.745410] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-389d1938-5433-4e00-aa35-1950b659f408 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.752021] env[62952]: DEBUG oslo_vmware.api [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1068.752021] env[62952]: value = "task-1263808" [ 1068.752021] env[62952]: _type = "Task" [ 1068.752021] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.760305] env[62952]: DEBUG oslo_vmware.api [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263808, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.966359] env[62952]: DEBUG oslo_vmware.api [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263805, 'name': ReconfigVM_Task, 'duration_secs': 0.16245} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.966676] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271963', 'volume_id': '1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e', 'name': 'volume-1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e44914cd-d8ae-4f43-9207-32cacf6d6553', 'attached_at': '', 'detached_at': '', 'volume_id': '1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e', 'serial': '1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1069.028473] env[62952]: DEBUG nova.compute.manager [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1069.040923] env[62952]: DEBUG oslo_concurrency.lockutils [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "1941365e-d562-4d99-89eb-226a07e52071" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.041190] env[62952]: DEBUG oslo_concurrency.lockutils [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "1941365e-d562-4d99-89eb-226a07e52071" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.041424] env[62952]: INFO nova.compute.manager [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Attaching volume 7ee58197-e579-4b82-9820-78f9798fc80e to /dev/sdb [ 1069.077146] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904b838d-669d-458c-8ba0-4e043faec1ac {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.087601] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b30e494c-d702-4feb-bc67-1a1894106653 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.093386] env[62952]: DEBUG oslo_vmware.api [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263806, 'name': PowerOffVM_Task, 'duration_secs': 0.188839} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.094011] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1069.094266] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance '6e289c20-8dd2-4680-b816-9bbe82fe7aec' progress to 17 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1069.104302] env[62952]: DEBUG nova.virt.block_device [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Updating existing volume attachment record: 6f299611-1ad8-4bb1-9f87-3d4e2bcf2cbe {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1069.223831] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': task-1263807, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069165} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.223831] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1069.224400] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5422335-5869-4594-aeea-32bdce48942d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.246557] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 1a02fba4-aad0-4bd4-8183-a7f08d887a5a/1a02fba4-aad0-4bd4-8183-a7f08d887a5a.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1069.247566] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.247812] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.248035] env[62952]: DEBUG nova.objects.instance [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lazy-loading 'resources' on Instance uuid af261b32-457a-44a1-bab4-b62654b0be1f {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.249067] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-013c6c05-8ed7-4450-8c45-aa3b0eea2d96 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.275242] env[62952]: DEBUG oslo_vmware.api [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263808, 'name': PowerOnVM_Task, 'duration_secs': 0.354561} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.276518] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1069.276725] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87e6f239-c9c3-4747-bf6e-e6a54817f969 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Updating instance 'e9e690b8-2fae-401d-aaad-0acef6e88ce3' progress to 100 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1069.280735] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Waiting for the task: (returnval){ [ 1069.280735] env[62952]: value = "task-1263810" [ 1069.280735] env[62952]: _type = "Task" [ 1069.280735] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.291762] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': task-1263810, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.442198] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f2ca067-fae1-4567-9e4f-8025836ea042 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.451884] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4dbf765-57a2-41e4-9eae-7741363bb50f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.486859] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31de8a7c-72c0-4f60-ab04-5b5dd5fed455 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.493985] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c22dd6c2-fe24-4d4c-8e8b-01ad4a86596a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.507162] env[62952]: DEBUG nova.compute.provider_tree [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1069.549157] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.600798] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1069.600798] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1069.600798] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1069.601079] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1069.601674] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1069.601674] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1069.601925] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1069.602254] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1069.602578] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1069.603168] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1069.603168] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1069.608324] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c402bcb5-287e-471a-aba5-387afac95eba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.623990] env[62952]: DEBUG oslo_vmware.api [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1069.623990] env[62952]: value = "task-1263813" [ 1069.623990] env[62952]: _type = "Task" [ 1069.623990] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.632210] env[62952]: DEBUG oslo_vmware.api [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263813, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.793460] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': task-1263810, 'name': ReconfigVM_Task, 'duration_secs': 0.405684} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.793703] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 1a02fba4-aad0-4bd4-8183-a7f08d887a5a/1a02fba4-aad0-4bd4-8183-a7f08d887a5a.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1069.794799] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-481538f7-1ee4-437a-ba54-a9e522d2404b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.800359] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Waiting for the task: (returnval){ [ 1069.800359] env[62952]: value = "task-1263814" [ 1069.800359] env[62952]: _type = "Task" [ 1069.800359] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.808160] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': task-1263814, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.010403] env[62952]: DEBUG nova.scheduler.client.report [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1070.020980] env[62952]: DEBUG nova.objects.instance [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lazy-loading 'flavor' on Instance uuid e44914cd-d8ae-4f43-9207-32cacf6d6553 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.133521] env[62952]: DEBUG oslo_vmware.api [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263813, 'name': ReconfigVM_Task, 'duration_secs': 0.190481} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.133954] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance '6e289c20-8dd2-4680-b816-9bbe82fe7aec' progress to 33 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1070.310838] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': task-1263814, 'name': Rename_Task, 'duration_secs': 0.284535} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.312314] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1070.316114] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d6e05996-0ddd-49e1-a37a-e58594499a90 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.323991] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Waiting for the task: (returnval){ [ 1070.323991] env[62952]: value = "task-1263815" [ 1070.323991] env[62952]: _type = "Task" [ 1070.323991] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.333528] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': task-1263815, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.515525] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.267s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.517953] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.969s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.519533] env[62952]: INFO nova.compute.claims [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1070.525710] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b3cb671-7a90-4fab-a266-7530f5a2da91 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.279s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.539643] env[62952]: INFO nova.scheduler.client.report [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleted allocations for instance af261b32-457a-44a1-bab4-b62654b0be1f [ 1070.640915] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1070.641211] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1070.641393] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1070.641583] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1070.641733] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1070.641884] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1070.642102] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1070.642271] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1070.642446] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1070.642612] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1070.642787] env[62952]: DEBUG nova.virt.hardware [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1070.648386] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Reconfiguring VM instance instance-0000003c to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1070.649050] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f831222b-f84d-420c-a95c-a8925e9eb3e7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.670079] env[62952]: DEBUG oslo_vmware.api [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1070.670079] env[62952]: value = "task-1263816" [ 1070.670079] env[62952]: _type = "Task" [ 1070.670079] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.678205] env[62952]: DEBUG oslo_vmware.api [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263816, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.818666] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.818953] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.833860] env[62952]: DEBUG oslo_vmware.api [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': task-1263815, 'name': PowerOnVM_Task, 'duration_secs': 0.438764} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.834148] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1070.834604] env[62952]: INFO nova.compute.manager [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Took 7.32 seconds to spawn the instance on the hypervisor. [ 1070.834604] env[62952]: DEBUG nova.compute.manager [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1070.835291] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a9300c0-41eb-4aa6-bc2f-a32963f4b1d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.045895] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f5997392-7ee8-4b4f-a486-5b3c8aec2650 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "af261b32-457a-44a1-bab4-b62654b0be1f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.216s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.083793] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.083902] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.084106] env[62952]: DEBUG nova.compute.manager [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Going to confirm migration 2 {{(pid=62952) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1071.180662] env[62952]: DEBUG oslo_vmware.api [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263816, 'name': ReconfigVM_Task, 'duration_secs': 0.167072} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.180967] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Reconfigured VM instance instance-0000003c to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1071.181773] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6fb7bb3-21ca-4b39-8b49-54564b72d379 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.205938] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 6e289c20-8dd2-4680-b816-9bbe82fe7aec/6e289c20-8dd2-4680-b816-9bbe82fe7aec.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1071.206602] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc66d87e-bb0c-481b-819e-2e8c686e4a26 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.225642] env[62952]: DEBUG oslo_vmware.api [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1071.225642] env[62952]: value = "task-1263817" [ 1071.225642] env[62952]: _type = "Task" [ 1071.225642] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.233180] env[62952]: DEBUG oslo_vmware.api [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263817, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.322996] env[62952]: INFO nova.compute.manager [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Detaching volume 3b174d32-3a39-4759-ba6c-6ea837f89b2e [ 1071.353532] env[62952]: INFO nova.compute.manager [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Took 18.58 seconds to build instance. [ 1071.363107] env[62952]: INFO nova.virt.block_device [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Attempting to driver detach volume 3b174d32-3a39-4759-ba6c-6ea837f89b2e from mountpoint /dev/sdb [ 1071.363395] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1071.363663] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271961', 'volume_id': '3b174d32-3a39-4759-ba6c-6ea837f89b2e', 'name': 'volume-3b174d32-3a39-4759-ba6c-6ea837f89b2e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e44914cd-d8ae-4f43-9207-32cacf6d6553', 'attached_at': '', 'detached_at': '', 'volume_id': '3b174d32-3a39-4759-ba6c-6ea837f89b2e', 'serial': '3b174d32-3a39-4759-ba6c-6ea837f89b2e'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1071.366865] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bfc8a27-e5ef-4755-8be9-a6277965d6c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.392872] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f87be517-a444-4597-a4c4-6e8c93a26d9b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.400277] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef952137-4b74-40ab-9e85-ce7965397dd2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.424013] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a7906df-40f3-429d-9367-ef3564a9b326 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.439630] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] The volume has not been displaced from its original location: [datastore2] volume-3b174d32-3a39-4759-ba6c-6ea837f89b2e/volume-3b174d32-3a39-4759-ba6c-6ea837f89b2e.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1071.445060] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Reconfiguring VM instance instance-00000055 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1071.445422] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb156832-ba60-48d2-b532-4fcb67a8f6b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.464511] env[62952]: DEBUG oslo_vmware.api [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1071.464511] env[62952]: value = "task-1263819" [ 1071.464511] env[62952]: _type = "Task" [ 1071.464511] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.476386] env[62952]: DEBUG oslo_vmware.api [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263819, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.631046] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "refresh_cache-e9e690b8-2fae-401d-aaad-0acef6e88ce3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.631438] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "refresh_cache-e9e690b8-2fae-401d-aaad-0acef6e88ce3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.631438] env[62952]: DEBUG nova.network.neutron [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1071.631625] env[62952]: DEBUG nova.objects.instance [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lazy-loading 'info_cache' on Instance uuid e9e690b8-2fae-401d-aaad-0acef6e88ce3 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.687174] env[62952]: DEBUG oslo_concurrency.lockutils [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "8cda6eea-d488-4d73-a7ab-7a905037929d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.687354] env[62952]: DEBUG oslo_concurrency.lockutils [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "8cda6eea-d488-4d73-a7ab-7a905037929d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.704770] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8c153b-7c5b-49a0-bc8c-486028ae782d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.712517] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f2bb18-68f1-4737-a5b2-2dec78ecd52a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.746374] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d93b6411-9fba-42e0-9d98-63545be02e2e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.756638] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c34172c-c420-4dcd-ba66-982ab4f84195 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.760670] env[62952]: DEBUG oslo_vmware.api [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263817, 'name': ReconfigVM_Task, 'duration_secs': 0.297443} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.761046] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 6e289c20-8dd2-4680-b816-9bbe82fe7aec/6e289c20-8dd2-4680-b816-9bbe82fe7aec.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1071.761339] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance '6e289c20-8dd2-4680-b816-9bbe82fe7aec' progress to 50 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1071.775909] env[62952]: DEBUG nova.compute.provider_tree [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.856048] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3c987bce-eac8-4d74-86de-cb78ac19a23d tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Lock "1a02fba4-aad0-4bd4-8183-a7f08d887a5a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.093s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.974783] env[62952]: DEBUG oslo_vmware.api [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263819, 'name': ReconfigVM_Task, 'duration_secs': 0.348543} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.975085] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Reconfigured VM instance instance-00000055 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1071.979872] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b97b5ef2-79ed-43cc-982d-8d73fb3bcc09 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.994644] env[62952]: DEBUG oslo_vmware.api [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1071.994644] env[62952]: value = "task-1263820" [ 1071.994644] env[62952]: _type = "Task" [ 1071.994644] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.002088] env[62952]: DEBUG oslo_vmware.api [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263820, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.167846] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Acquiring lock "1a02fba4-aad0-4bd4-8183-a7f08d887a5a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.168220] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Lock "1a02fba4-aad0-4bd4-8183-a7f08d887a5a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.168445] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Acquiring lock "1a02fba4-aad0-4bd4-8183-a7f08d887a5a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.168635] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Lock "1a02fba4-aad0-4bd4-8183-a7f08d887a5a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.168815] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Lock "1a02fba4-aad0-4bd4-8183-a7f08d887a5a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.170941] env[62952]: INFO nova.compute.manager [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Terminating instance [ 1072.172708] env[62952]: DEBUG nova.compute.manager [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1072.172906] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1072.173733] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e9bd67b-8c44-4cf1-8449-199ad907d553 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.182082] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1072.182345] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-13baf1db-bdac-472d-be3f-c0c4307015a2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.189051] env[62952]: DEBUG oslo_vmware.api [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Waiting for the task: (returnval){ [ 1072.189051] env[62952]: value = "task-1263821" [ 1072.189051] env[62952]: _type = "Task" [ 1072.189051] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.189051] env[62952]: DEBUG nova.compute.manager [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1072.200343] env[62952]: DEBUG oslo_vmware.api [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': task-1263821, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.269026] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817f134b-8fc7-4dfc-bab0-2917ab548c15 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.288886] env[62952]: DEBUG nova.scheduler.client.report [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1072.294066] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b05725fe-37ef-4c0c-be52-c107ddb9d401 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.311441] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance '6e289c20-8dd2-4680-b816-9bbe82fe7aec' progress to 67 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1072.504254] env[62952]: DEBUG oslo_vmware.api [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263820, 'name': ReconfigVM_Task, 'duration_secs': 0.244369} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.504588] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271961', 'volume_id': '3b174d32-3a39-4759-ba6c-6ea837f89b2e', 'name': 'volume-3b174d32-3a39-4759-ba6c-6ea837f89b2e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e44914cd-d8ae-4f43-9207-32cacf6d6553', 'attached_at': '', 'detached_at': '', 'volume_id': '3b174d32-3a39-4759-ba6c-6ea837f89b2e', 'serial': '3b174d32-3a39-4759-ba6c-6ea837f89b2e'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1072.702076] env[62952]: DEBUG oslo_vmware.api [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': task-1263821, 'name': PowerOffVM_Task, 'duration_secs': 0.21352} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.704203] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1072.704380] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1072.704811] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80ab6f8d-678c-4671-8552-4928ce4095a7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.718647] env[62952]: DEBUG oslo_concurrency.lockutils [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.769329] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1072.769575] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1072.769767] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Deleting the datastore file [datastore1] 1a02fba4-aad0-4bd4-8183-a7f08d887a5a {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1072.770084] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b0f3ad7e-d3b3-4ffc-97fa-5e5f4e7c84a3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.776429] env[62952]: DEBUG oslo_vmware.api [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Waiting for the task: (returnval){ [ 1072.776429] env[62952]: value = "task-1263823" [ 1072.776429] env[62952]: _type = "Task" [ 1072.776429] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.784654] env[62952]: DEBUG oslo_vmware.api [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': task-1263823, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.797534] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.280s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.798108] env[62952]: DEBUG nova.compute.manager [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1072.800898] env[62952]: DEBUG oslo_concurrency.lockutils [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.082s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.804019] env[62952]: INFO nova.compute.claims [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1072.873981] env[62952]: DEBUG nova.network.neutron [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Port 5e2e9d1e-dce4-404e-bf41-1a7897be3758 binding to destination host cpu-1 is already ACTIVE {{(pid=62952) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1073.046483] env[62952]: DEBUG nova.objects.instance [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lazy-loading 'flavor' on Instance uuid e44914cd-d8ae-4f43-9207-32cacf6d6553 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1073.142348] env[62952]: DEBUG nova.network.neutron [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Updating instance_info_cache with network_info: [{"id": "cff4c81c-513a-42a4-b6f5-94b4367fa0ea", "address": "fa:16:3e:3a:e4:22", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcff4c81c-51", "ovs_interfaceid": "cff4c81c-513a-42a4-b6f5-94b4367fa0ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.287175] env[62952]: DEBUG oslo_vmware.api [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Task: {'id': task-1263823, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139371} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.287480] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1073.287651] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1073.287824] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1073.288011] env[62952]: INFO nova.compute.manager [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1073.288297] env[62952]: DEBUG oslo.service.loopingcall [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1073.288506] env[62952]: DEBUG nova.compute.manager [-] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1073.288602] env[62952]: DEBUG nova.network.neutron [-] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1073.306416] env[62952]: DEBUG nova.compute.utils [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1073.309802] env[62952]: DEBUG nova.compute.manager [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1073.309802] env[62952]: DEBUG nova.network.neutron [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1073.348902] env[62952]: DEBUG nova.policy [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7eadc05f193645e6b6e86e9cdbbbaa71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '076fb3c53420447caf047f8422435152', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1073.537737] env[62952]: DEBUG nova.compute.manager [req-128b8725-86a9-47b0-8e1d-fa92de2f4b77 req-9a06df44-ae39-465e-8391-59499cc32d5e service nova] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Received event network-vif-deleted-87c0fa2c-a6cf-4539-a9d4-3c4394c10068 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1073.537737] env[62952]: INFO nova.compute.manager [req-128b8725-86a9-47b0-8e1d-fa92de2f4b77 req-9a06df44-ae39-465e-8391-59499cc32d5e service nova] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Neutron deleted interface 87c0fa2c-a6cf-4539-a9d4-3c4394c10068; detaching it from the instance and deleting it from the info cache [ 1073.537907] env[62952]: DEBUG nova.network.neutron [req-128b8725-86a9-47b0-8e1d-fa92de2f4b77 req-9a06df44-ae39-465e-8391-59499cc32d5e service nova] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.612237] env[62952]: DEBUG nova.network.neutron [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Successfully created port: c764e89c-8999-4422-ab55-e64bf2dee40b {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1073.644879] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "refresh_cache-e9e690b8-2fae-401d-aaad-0acef6e88ce3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.645101] env[62952]: DEBUG nova.objects.instance [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lazy-loading 'migration_context' on Instance uuid e9e690b8-2fae-401d-aaad-0acef6e88ce3 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1073.650590] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1073.650590] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271968', 'volume_id': '7ee58197-e579-4b82-9820-78f9798fc80e', 'name': 'volume-7ee58197-e579-4b82-9820-78f9798fc80e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1941365e-d562-4d99-89eb-226a07e52071', 'attached_at': '', 'detached_at': '', 'volume_id': '7ee58197-e579-4b82-9820-78f9798fc80e', 'serial': '7ee58197-e579-4b82-9820-78f9798fc80e'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1073.651512] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7653273-99dd-405f-b7c6-7c09f2c065b7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.668175] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55820af5-091b-4a54-bac2-4fc15af57502 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.696850] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] volume-7ee58197-e579-4b82-9820-78f9798fc80e/volume-7ee58197-e579-4b82-9820-78f9798fc80e.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1073.697183] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7376d4d7-3cf7-4b64-bba1-8fed576f5838 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.715834] env[62952]: DEBUG oslo_vmware.api [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1073.715834] env[62952]: value = "task-1263824" [ 1073.715834] env[62952]: _type = "Task" [ 1073.715834] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.724958] env[62952]: DEBUG oslo_vmware.api [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263824, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.810539] env[62952]: DEBUG nova.compute.manager [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1073.901053] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.901053] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.901053] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.999422] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5df358-396a-48e7-9389-748731899ff8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.006890] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b680d5ef-1656-42d4-964f-c96c8fea65d3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.009933] env[62952]: DEBUG nova.network.neutron [-] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.041030] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f2d8993-44f7-44b9-8fc5-03e965567d43 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.044254] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cd3f4e9f-d96f-4bbf-8cbe-a90db4181a2e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.052179] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e64bc0-bf12-46a3-943e-ace2908a34c0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.056306] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2643fd26-9682-4a09-a5a8-c3f7c8182aa4 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.237s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.060827] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a533c9e9-ce0b-46dd-93f3-d7a3a38ce5f1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.081514] env[62952]: DEBUG nova.compute.provider_tree [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1074.092411] env[62952]: DEBUG nova.compute.manager [req-128b8725-86a9-47b0-8e1d-fa92de2f4b77 req-9a06df44-ae39-465e-8391-59499cc32d5e service nova] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Detach interface failed, port_id=87c0fa2c-a6cf-4539-a9d4-3c4394c10068, reason: Instance 1a02fba4-aad0-4bd4-8183-a7f08d887a5a could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1074.111112] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.111112] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.147858] env[62952]: DEBUG nova.objects.base [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1074.148772] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bcff7f6-d48a-4844-b7df-a5a941622591 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.169168] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22b19342-1774-4ce2-964c-000122493577 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.174795] env[62952]: DEBUG oslo_vmware.api [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1074.174795] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52eab6e7-5227-8cb2-c1cc-d3eb5ff7d6f9" [ 1074.174795] env[62952]: _type = "Task" [ 1074.174795] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.183041] env[62952]: DEBUG oslo_vmware.api [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52eab6e7-5227-8cb2-c1cc-d3eb5ff7d6f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.226491] env[62952]: DEBUG oslo_vmware.api [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263824, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.512312] env[62952]: INFO nova.compute.manager [-] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Took 1.22 seconds to deallocate network for instance. [ 1074.593845] env[62952]: DEBUG nova.scheduler.client.report [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1074.612757] env[62952]: INFO nova.compute.manager [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Detaching volume 1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e [ 1074.643774] env[62952]: INFO nova.virt.block_device [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Attempting to driver detach volume 1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e from mountpoint /dev/sdc [ 1074.644033] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1074.644233] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271963', 'volume_id': '1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e', 'name': 'volume-1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e44914cd-d8ae-4f43-9207-32cacf6d6553', 'attached_at': '', 'detached_at': '', 'volume_id': '1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e', 'serial': '1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1074.645109] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97408a74-d178-4b4d-b4e4-682a4ea9a4d6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.666424] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1963b8d-2edf-4171-b9a8-3d58dadfca4f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.673427] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2384dd6-383c-4f9c-ba5c-2344ada7e97e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.698828] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b4cfec-c50b-41f4-9746-69083e50f63f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.701269] env[62952]: DEBUG oslo_vmware.api [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52eab6e7-5227-8cb2-c1cc-d3eb5ff7d6f9, 'name': SearchDatastore_Task, 'duration_secs': 0.007} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.701905] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.717151] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] The volume has not been displaced from its original location: [datastore1] volume-1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e/volume-1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1074.720600] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Reconfiguring VM instance instance-00000055 to detach disk 2002 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1074.720695] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2747629b-b4bb-449c-8114-a29a294d2306 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.741358] env[62952]: DEBUG oslo_vmware.api [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263824, 'name': ReconfigVM_Task, 'duration_secs': 0.522437} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.743164] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Reconfigured VM instance instance-00000059 to attach disk [datastore2] volume-7ee58197-e579-4b82-9820-78f9798fc80e/volume-7ee58197-e579-4b82-9820-78f9798fc80e.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1074.746706] env[62952]: DEBUG oslo_vmware.api [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1074.746706] env[62952]: value = "task-1263825" [ 1074.746706] env[62952]: _type = "Task" [ 1074.746706] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.746907] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c03cfb2d-b1e9-4908-a4e4-ca0fa5675d39 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.765529] env[62952]: DEBUG oslo_vmware.api [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.766913] env[62952]: DEBUG oslo_vmware.api [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1074.766913] env[62952]: value = "task-1263826" [ 1074.766913] env[62952]: _type = "Task" [ 1074.766913] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.775257] env[62952]: DEBUG oslo_vmware.api [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263826, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.822048] env[62952]: DEBUG nova.compute.manager [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1074.848543] env[62952]: DEBUG nova.virt.hardware [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1074.848811] env[62952]: DEBUG nova.virt.hardware [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1074.848974] env[62952]: DEBUG nova.virt.hardware [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1074.849185] env[62952]: DEBUG nova.virt.hardware [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1074.849336] env[62952]: DEBUG nova.virt.hardware [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1074.849488] env[62952]: DEBUG nova.virt.hardware [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1074.849702] env[62952]: DEBUG nova.virt.hardware [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1074.849868] env[62952]: DEBUG nova.virt.hardware [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1074.850049] env[62952]: DEBUG nova.virt.hardware [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1074.850223] env[62952]: DEBUG nova.virt.hardware [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1074.850493] env[62952]: DEBUG nova.virt.hardware [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1074.851355] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e8d284-ab97-4f22-b4d1-317b4f248b87 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.861424] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c75db1ce-f51d-43ac-9334-8b4dbf24d547 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.953176] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.953373] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.954268] env[62952]: DEBUG nova.network.neutron [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1075.012744] env[62952]: DEBUG nova.compute.manager [req-fad5bb59-8b42-414c-8bee-fe1fd0326fb5 req-5e177737-991c-4be8-9727-f48e5d73a102 service nova] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Received event network-vif-plugged-c764e89c-8999-4422-ab55-e64bf2dee40b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1075.012946] env[62952]: DEBUG oslo_concurrency.lockutils [req-fad5bb59-8b42-414c-8bee-fe1fd0326fb5 req-5e177737-991c-4be8-9727-f48e5d73a102 service nova] Acquiring lock "f28d0043-236e-4df4-8478-e4b1befb6a7d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.013568] env[62952]: DEBUG oslo_concurrency.lockutils [req-fad5bb59-8b42-414c-8bee-fe1fd0326fb5 req-5e177737-991c-4be8-9727-f48e5d73a102 service nova] Lock "f28d0043-236e-4df4-8478-e4b1befb6a7d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.013801] env[62952]: DEBUG oslo_concurrency.lockutils [req-fad5bb59-8b42-414c-8bee-fe1fd0326fb5 req-5e177737-991c-4be8-9727-f48e5d73a102 service nova] Lock "f28d0043-236e-4df4-8478-e4b1befb6a7d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.014013] env[62952]: DEBUG nova.compute.manager [req-fad5bb59-8b42-414c-8bee-fe1fd0326fb5 req-5e177737-991c-4be8-9727-f48e5d73a102 service nova] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] No waiting events found dispatching network-vif-plugged-c764e89c-8999-4422-ab55-e64bf2dee40b {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1075.014262] env[62952]: WARNING nova.compute.manager [req-fad5bb59-8b42-414c-8bee-fe1fd0326fb5 req-5e177737-991c-4be8-9727-f48e5d73a102 service nova] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Received unexpected event network-vif-plugged-c764e89c-8999-4422-ab55-e64bf2dee40b for instance with vm_state building and task_state spawning. [ 1075.017860] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.099109] env[62952]: DEBUG nova.network.neutron [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Successfully updated port: c764e89c-8999-4422-ab55-e64bf2dee40b {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1075.101499] env[62952]: DEBUG oslo_concurrency.lockutils [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.301s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.102092] env[62952]: DEBUG nova.compute.manager [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1075.105408] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.404s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.267177] env[62952]: DEBUG oslo_vmware.api [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263825, 'name': ReconfigVM_Task, 'duration_secs': 0.243915} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.267551] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Reconfigured VM instance instance-00000055 to detach disk 2002 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1075.274865] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8521a048-1826-4714-b2de-5afe3de95d58 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.289493] env[62952]: DEBUG oslo_vmware.api [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263826, 'name': ReconfigVM_Task, 'duration_secs': 0.136763} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.290669] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271968', 'volume_id': '7ee58197-e579-4b82-9820-78f9798fc80e', 'name': 'volume-7ee58197-e579-4b82-9820-78f9798fc80e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1941365e-d562-4d99-89eb-226a07e52071', 'attached_at': '', 'detached_at': '', 'volume_id': '7ee58197-e579-4b82-9820-78f9798fc80e', 'serial': '7ee58197-e579-4b82-9820-78f9798fc80e'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1075.292066] env[62952]: DEBUG oslo_vmware.api [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1075.292066] env[62952]: value = "task-1263827" [ 1075.292066] env[62952]: _type = "Task" [ 1075.292066] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.299829] env[62952]: DEBUG oslo_vmware.api [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263827, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.606780] env[62952]: DEBUG nova.compute.utils [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1075.608305] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "refresh_cache-f28d0043-236e-4df4-8478-e4b1befb6a7d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.608433] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "refresh_cache-f28d0043-236e-4df4-8478-e4b1befb6a7d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.608573] env[62952]: DEBUG nova.network.neutron [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1075.609731] env[62952]: DEBUG nova.compute.manager [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1075.609893] env[62952]: DEBUG nova.network.neutron [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1075.654614] env[62952]: DEBUG nova.policy [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '469eae9db3d14e9f90988e3093514e69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff928791a3cf496cb8db2c2a6c47a6a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1075.660665] env[62952]: DEBUG nova.network.neutron [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance_info_cache with network_info: [{"id": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "address": "fa:16:3e:eb:c2:8d", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e2e9d1e-dc", "ovs_interfaceid": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.778712] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b12dd2fa-0fd4-4991-8820-965fbd32be23 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.786224] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4cc9af-5a52-4bff-9223-1b1ab8702a7d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.823804] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20f64dfd-1e23-498e-a757-96f5eb4dd121 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.833968] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e040523-0b36-4888-8b1c-7d6cbe780127 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.837843] env[62952]: DEBUG oslo_vmware.api [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263827, 'name': ReconfigVM_Task, 'duration_secs': 0.143601} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.838165] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271963', 'volume_id': '1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e', 'name': 'volume-1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e44914cd-d8ae-4f43-9207-32cacf6d6553', 'attached_at': '', 'detached_at': '', 'volume_id': '1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e', 'serial': '1d5774f8-5ddd-48e3-8d93-fd54e9f81f2e'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1075.849486] env[62952]: DEBUG nova.compute.provider_tree [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1075.979676] env[62952]: DEBUG nova.network.neutron [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Successfully created port: 116c22bf-b489-4039-9b75-598fe7be83dc {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1076.111326] env[62952]: DEBUG nova.compute.manager [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1076.147497] env[62952]: DEBUG nova.network.neutron [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1076.166140] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.296509] env[62952]: DEBUG nova.network.neutron [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Updating instance_info_cache with network_info: [{"id": "c764e89c-8999-4422-ab55-e64bf2dee40b", "address": "fa:16:3e:95:67:26", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc764e89c-89", "ovs_interfaceid": "c764e89c-8999-4422-ab55-e64bf2dee40b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.355445] env[62952]: DEBUG nova.scheduler.client.report [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1076.359320] env[62952]: DEBUG nova.objects.instance [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lazy-loading 'flavor' on Instance uuid 1941365e-d562-4d99-89eb-226a07e52071 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1076.389274] env[62952]: DEBUG nova.objects.instance [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lazy-loading 'flavor' on Instance uuid e44914cd-d8ae-4f43-9207-32cacf6d6553 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1076.689540] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d80f28-cc57-4718-a810-d899d1e8e162 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.708499] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-976a98a5-f5df-4001-8c68-5f63e01d8966 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.715341] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance '6e289c20-8dd2-4680-b816-9bbe82fe7aec' progress to 83 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1076.799407] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "refresh_cache-f28d0043-236e-4df4-8478-e4b1befb6a7d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.799742] env[62952]: DEBUG nova.compute.manager [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Instance network_info: |[{"id": "c764e89c-8999-4422-ab55-e64bf2dee40b", "address": "fa:16:3e:95:67:26", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc764e89c-89", "ovs_interfaceid": "c764e89c-8999-4422-ab55-e64bf2dee40b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1076.800187] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:67:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '365ac5b1-6d83-4dfe-887f-60574d7f6124', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c764e89c-8999-4422-ab55-e64bf2dee40b', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1076.808820] env[62952]: DEBUG oslo.service.loopingcall [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1076.809064] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1076.809300] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ef47d40-471e-40be-b1cf-5a8d57a61a30 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.829573] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1076.829573] env[62952]: value = "task-1263828" [ 1076.829573] env[62952]: _type = "Task" [ 1076.829573] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.832684] env[62952]: INFO nova.compute.manager [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Rescuing [ 1076.832924] env[62952]: DEBUG oslo_concurrency.lockutils [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.833135] env[62952]: DEBUG oslo_concurrency.lockutils [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.833269] env[62952]: DEBUG nova.network.neutron [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1076.838625] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263828, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.866695] env[62952]: DEBUG oslo_concurrency.lockutils [None req-df4f127c-8c42-4702-ba61-b8ebc44671df tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "1941365e-d562-4d99-89eb-226a07e52071" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.825s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.103887] env[62952]: DEBUG nova.compute.manager [req-37ec9ac2-c06e-44dc-ab1a-599bcad8d843 req-a1a223aa-1e51-4068-8277-b46e0e847e6f service nova] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Received event network-changed-c764e89c-8999-4422-ab55-e64bf2dee40b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1077.103887] env[62952]: DEBUG nova.compute.manager [req-37ec9ac2-c06e-44dc-ab1a-599bcad8d843 req-a1a223aa-1e51-4068-8277-b46e0e847e6f service nova] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Refreshing instance network info cache due to event network-changed-c764e89c-8999-4422-ab55-e64bf2dee40b. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1077.104010] env[62952]: DEBUG oslo_concurrency.lockutils [req-37ec9ac2-c06e-44dc-ab1a-599bcad8d843 req-a1a223aa-1e51-4068-8277-b46e0e847e6f service nova] Acquiring lock "refresh_cache-f28d0043-236e-4df4-8478-e4b1befb6a7d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.104164] env[62952]: DEBUG oslo_concurrency.lockutils [req-37ec9ac2-c06e-44dc-ab1a-599bcad8d843 req-a1a223aa-1e51-4068-8277-b46e0e847e6f service nova] Acquired lock "refresh_cache-f28d0043-236e-4df4-8478-e4b1befb6a7d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.104342] env[62952]: DEBUG nova.network.neutron [req-37ec9ac2-c06e-44dc-ab1a-599bcad8d843 req-a1a223aa-1e51-4068-8277-b46e0e847e6f service nova] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Refreshing network info cache for port c764e89c-8999-4422-ab55-e64bf2dee40b {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1077.120227] env[62952]: DEBUG nova.compute.manager [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1077.144911] env[62952]: DEBUG nova.virt.hardware [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1077.145189] env[62952]: DEBUG nova.virt.hardware [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1077.145353] env[62952]: DEBUG nova.virt.hardware [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1077.145541] env[62952]: DEBUG nova.virt.hardware [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1077.145724] env[62952]: DEBUG nova.virt.hardware [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1077.145881] env[62952]: DEBUG nova.virt.hardware [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1077.146111] env[62952]: DEBUG nova.virt.hardware [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1077.146287] env[62952]: DEBUG nova.virt.hardware [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1077.146461] env[62952]: DEBUG nova.virt.hardware [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1077.146628] env[62952]: DEBUG nova.virt.hardware [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1077.146802] env[62952]: DEBUG nova.virt.hardware [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1077.147678] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fecdea08-14f3-4a6c-96ad-2344f98059e3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.155222] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eedc4712-6d47-4b58-aafb-7ac2c74aa55f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.223299] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1077.223577] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2c993874-e7c7-453f-acb5-5673b7d825e7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.230615] env[62952]: DEBUG oslo_vmware.api [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1077.230615] env[62952]: value = "task-1263829" [ 1077.230615] env[62952]: _type = "Task" [ 1077.230615] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.240456] env[62952]: DEBUG oslo_vmware.api [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263829, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.340466] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263828, 'name': CreateVM_Task, 'duration_secs': 0.344509} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.340946] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1077.341611] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.341781] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.342108] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1077.342376] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc6c16e5-bcc0-46b4-b8b6-6f7ed5291396 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.346877] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1077.346877] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]529447b1-b7b8-4d9f-cdf8-3849ccbf2eed" [ 1077.346877] env[62952]: _type = "Task" [ 1077.346877] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.353186] env[62952]: DEBUG nova.compute.manager [req-8bb78ffe-e3c0-4119-a81d-2c4a529e3f6c req-f622a650-260f-4280-b8d5-4c694d54a230 service nova] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Received event network-vif-plugged-116c22bf-b489-4039-9b75-598fe7be83dc {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1077.353395] env[62952]: DEBUG oslo_concurrency.lockutils [req-8bb78ffe-e3c0-4119-a81d-2c4a529e3f6c req-f622a650-260f-4280-b8d5-4c694d54a230 service nova] Acquiring lock "8cda6eea-d488-4d73-a7ab-7a905037929d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.353612] env[62952]: DEBUG oslo_concurrency.lockutils [req-8bb78ffe-e3c0-4119-a81d-2c4a529e3f6c req-f622a650-260f-4280-b8d5-4c694d54a230 service nova] Lock "8cda6eea-d488-4d73-a7ab-7a905037929d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.353870] env[62952]: DEBUG oslo_concurrency.lockutils [req-8bb78ffe-e3c0-4119-a81d-2c4a529e3f6c req-f622a650-260f-4280-b8d5-4c694d54a230 service nova] Lock "8cda6eea-d488-4d73-a7ab-7a905037929d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.354070] env[62952]: DEBUG nova.compute.manager [req-8bb78ffe-e3c0-4119-a81d-2c4a529e3f6c req-f622a650-260f-4280-b8d5-4c694d54a230 service nova] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] No waiting events found dispatching network-vif-plugged-116c22bf-b489-4039-9b75-598fe7be83dc {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1077.354247] env[62952]: WARNING nova.compute.manager [req-8bb78ffe-e3c0-4119-a81d-2c4a529e3f6c req-f622a650-260f-4280-b8d5-4c694d54a230 service nova] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Received unexpected event network-vif-plugged-116c22bf-b489-4039-9b75-598fe7be83dc for instance with vm_state building and task_state spawning. [ 1077.357666] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]529447b1-b7b8-4d9f-cdf8-3849ccbf2eed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.368062] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.263s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.371436] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.353s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.371747] env[62952]: DEBUG nova.objects.instance [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Lazy-loading 'resources' on Instance uuid 1a02fba4-aad0-4bd4-8183-a7f08d887a5a {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1077.395604] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a3336315-c824-411d-b2d0-576d8ef92ca6 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.285s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.571942] env[62952]: DEBUG nova.network.neutron [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Updating instance_info_cache with network_info: [{"id": "e1c235af-9501-47c0-a83a-ba40ea7fa8c1", "address": "fa:16:3e:34:07:9b", "network": {"id": "827a8171-6759-4434-a6f1-437aad881c81", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1902486970-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091cdeb6da48492bb02f93822a45c9bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11b669be-fb26-4ef8-bdb6-c77ab9d06daf", "external-id": "nsx-vlan-transportzone-633", "segmentation_id": 633, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1c235af-95", "ovs_interfaceid": "e1c235af-9501-47c0-a83a-ba40ea7fa8c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.741144] env[62952]: DEBUG oslo_vmware.api [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263829, 'name': PowerOnVM_Task, 'duration_secs': 0.389233} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.741449] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1077.741572] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1c5a90fb-b75c-419d-97b6-94d1712979c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance '6e289c20-8dd2-4680-b816-9bbe82fe7aec' progress to 100 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1077.824808] env[62952]: DEBUG nova.network.neutron [req-37ec9ac2-c06e-44dc-ab1a-599bcad8d843 req-a1a223aa-1e51-4068-8277-b46e0e847e6f service nova] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Updated VIF entry in instance network info cache for port c764e89c-8999-4422-ab55-e64bf2dee40b. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1077.825312] env[62952]: DEBUG nova.network.neutron [req-37ec9ac2-c06e-44dc-ab1a-599bcad8d843 req-a1a223aa-1e51-4068-8277-b46e0e847e6f service nova] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Updating instance_info_cache with network_info: [{"id": "c764e89c-8999-4422-ab55-e64bf2dee40b", "address": "fa:16:3e:95:67:26", "network": {"id": "8822b1cd-d55c-4980-ae54-037637a75220", "bridge": "br-int", "label": "tempest-ServersTestJSON-351793537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076fb3c53420447caf047f8422435152", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc764e89c-89", "ovs_interfaceid": "c764e89c-8999-4422-ab55-e64bf2dee40b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.859016] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]529447b1-b7b8-4d9f-cdf8-3849ccbf2eed, 'name': SearchDatastore_Task, 'duration_secs': 0.009942} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.859016] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.859016] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1077.859016] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.859016] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.859016] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1077.859016] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-056fc34a-e2be-4fce-bd92-382858a4796e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.866065] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1077.866256] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1077.866971] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dad2a4d6-ae64-401e-bbab-953390f59aa2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.873141] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1077.873141] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e03c87-7937-5e06-f84f-a5888bb8927d" [ 1077.873141] env[62952]: _type = "Task" [ 1077.873141] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.886619] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e03c87-7937-5e06-f84f-a5888bb8927d, 'name': SearchDatastore_Task, 'duration_secs': 0.008395} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.889683] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73c27514-78c8-426d-b8ac-1cf8a8514a9c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.894426] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1077.894426] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52dfe238-225e-72f9-b03a-db22ed440ca0" [ 1077.894426] env[62952]: _type = "Task" [ 1077.894426] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.901836] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52dfe238-225e-72f9-b03a-db22ed440ca0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.926300] env[62952]: DEBUG nova.network.neutron [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Successfully updated port: 116c22bf-b489-4039-9b75-598fe7be83dc {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1077.951116] env[62952]: INFO nova.scheduler.client.report [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleted allocation for migration fdd5dcdb-d6fb-4c12-a4dd-cbd88d9a7609 [ 1077.983341] env[62952]: DEBUG nova.compute.manager [req-1582671a-b91b-4e7d-948c-4941f056d91d req-7628533f-4d20-4ab2-a52e-cda2467b56df service nova] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Received event network-changed-116c22bf-b489-4039-9b75-598fe7be83dc {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1077.983545] env[62952]: DEBUG nova.compute.manager [req-1582671a-b91b-4e7d-948c-4941f056d91d req-7628533f-4d20-4ab2-a52e-cda2467b56df service nova] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Refreshing instance network info cache due to event network-changed-116c22bf-b489-4039-9b75-598fe7be83dc. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1077.983765] env[62952]: DEBUG oslo_concurrency.lockutils [req-1582671a-b91b-4e7d-948c-4941f056d91d req-7628533f-4d20-4ab2-a52e-cda2467b56df service nova] Acquiring lock "refresh_cache-8cda6eea-d488-4d73-a7ab-7a905037929d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.983919] env[62952]: DEBUG oslo_concurrency.lockutils [req-1582671a-b91b-4e7d-948c-4941f056d91d req-7628533f-4d20-4ab2-a52e-cda2467b56df service nova] Acquired lock "refresh_cache-8cda6eea-d488-4d73-a7ab-7a905037929d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.984166] env[62952]: DEBUG nova.network.neutron [req-1582671a-b91b-4e7d-948c-4941f056d91d req-7628533f-4d20-4ab2-a52e-cda2467b56df service nova] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Refreshing network info cache for port 116c22bf-b489-4039-9b75-598fe7be83dc {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1078.074357] env[62952]: DEBUG oslo_concurrency.lockutils [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Releasing lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.078685] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e2c3c9-5e4e-453a-ae6b-4eeb67ece760 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.086292] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0acd7c-fbf4-432f-91cc-cf35acc5fd83 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.119749] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d8569b5-cc24-4b24-a0e1-9552f43fa531 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.127420] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acdac9d6-386a-495f-ad22-f055610e0705 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.144130] env[62952]: DEBUG nova.compute.provider_tree [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.327880] env[62952]: DEBUG oslo_concurrency.lockutils [req-37ec9ac2-c06e-44dc-ab1a-599bcad8d843 req-a1a223aa-1e51-4068-8277-b46e0e847e6f service nova] Releasing lock "refresh_cache-f28d0043-236e-4df4-8478-e4b1befb6a7d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.406476] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52dfe238-225e-72f9-b03a-db22ed440ca0, 'name': SearchDatastore_Task, 'duration_secs': 0.009079} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.406748] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.407046] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] f28d0043-236e-4df4-8478-e4b1befb6a7d/f28d0043-236e-4df4-8478-e4b1befb6a7d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1078.407292] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ed4a9419-1768-44c3-9017-d8dcc8721182 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.413920] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1078.413920] env[62952]: value = "task-1263830" [ 1078.413920] env[62952]: _type = "Task" [ 1078.413920] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.421763] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263830, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.430424] env[62952]: DEBUG oslo_concurrency.lockutils [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "refresh_cache-8cda6eea-d488-4d73-a7ab-7a905037929d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.462754] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8ee84bb9-32e9-441b-afa4-cdcb61dfb446 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.379s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.519289] env[62952]: DEBUG nova.network.neutron [req-1582671a-b91b-4e7d-948c-4941f056d91d req-7628533f-4d20-4ab2-a52e-cda2467b56df service nova] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1078.532259] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.532558] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.532800] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.533030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.533239] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.535531] env[62952]: INFO nova.compute.manager [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Terminating instance [ 1078.537466] env[62952]: DEBUG nova.compute.manager [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1078.537926] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1078.538593] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-febcc875-2632-4bd6-ae57-a83785c5377a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.546723] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1078.546991] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bb6c41a0-3927-4c49-b2d1-e5190b3e424f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.548671] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.548928] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.549173] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "e44914cd-d8ae-4f43-9207-32cacf6d6553-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.549385] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.549578] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.551780] env[62952]: INFO nova.compute.manager [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Terminating instance [ 1078.553522] env[62952]: DEBUG nova.compute.manager [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1078.553729] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1078.554540] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b74ae4-f35e-4953-beb3-20cc2856e28b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.558676] env[62952]: DEBUG oslo_vmware.api [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1078.558676] env[62952]: value = "task-1263831" [ 1078.558676] env[62952]: _type = "Task" [ 1078.558676] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.564179] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1078.564730] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a4172cc-5e1a-48e3-b1ad-c491cce86114 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.569854] env[62952]: DEBUG oslo_vmware.api [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263831, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.571081] env[62952]: DEBUG oslo_vmware.api [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1078.571081] env[62952]: value = "task-1263832" [ 1078.571081] env[62952]: _type = "Task" [ 1078.571081] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.580484] env[62952]: DEBUG oslo_vmware.api [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263832, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.614966] env[62952]: DEBUG nova.network.neutron [req-1582671a-b91b-4e7d-948c-4941f056d91d req-7628533f-4d20-4ab2-a52e-cda2467b56df service nova] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.640101] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1078.640457] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba7cb613-593a-429b-9337-f9b3fcd82a6e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.647577] env[62952]: DEBUG nova.scheduler.client.report [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1078.652582] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1078.652582] env[62952]: value = "task-1263833" [ 1078.652582] env[62952]: _type = "Task" [ 1078.652582] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.663049] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263833, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.926353] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263830, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.069269] env[62952]: DEBUG oslo_vmware.api [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263831, 'name': PowerOffVM_Task, 'duration_secs': 0.307404} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.069594] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1079.069776] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1079.070065] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-45686027-870d-4293-be33-83312b4883d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.081276] env[62952]: DEBUG oslo_vmware.api [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263832, 'name': PowerOffVM_Task, 'duration_secs': 0.29236} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.081577] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1079.081713] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1079.082310] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-37f50594-6b53-44ce-9957-c02ca3203fa5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.117906] env[62952]: DEBUG oslo_concurrency.lockutils [req-1582671a-b91b-4e7d-948c-4941f056d91d req-7628533f-4d20-4ab2-a52e-cda2467b56df service nova] Releasing lock "refresh_cache-8cda6eea-d488-4d73-a7ab-7a905037929d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.118392] env[62952]: DEBUG oslo_concurrency.lockutils [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "refresh_cache-8cda6eea-d488-4d73-a7ab-7a905037929d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.118587] env[62952]: DEBUG nova.network.neutron [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1079.132871] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "126f57b5-be03-4164-b92f-3c2e96dd9171" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.133127] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "126f57b5-be03-4164-b92f-3c2e96dd9171" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.154525] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.783s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.163420] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1079.163676] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1079.163863] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleting the datastore file [datastore1] e9e690b8-2fae-401d-aaad-0acef6e88ce3 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1079.164161] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bf1a68e4-5ae7-4c2c-9fd8-57361bd4a18c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.167723] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1079.167994] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1079.168220] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Deleting the datastore file [datastore2] e44914cd-d8ae-4f43-9207-32cacf6d6553 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1079.171071] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b97f93b2-ee28-4bd9-9899-89a6ec09b779 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.177149] env[62952]: DEBUG oslo_vmware.api [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1079.177149] env[62952]: value = "task-1263836" [ 1079.177149] env[62952]: _type = "Task" [ 1079.177149] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.177351] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263833, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.178706] env[62952]: INFO nova.scheduler.client.report [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Deleted allocations for instance 1a02fba4-aad0-4bd4-8183-a7f08d887a5a [ 1079.185467] env[62952]: DEBUG oslo_vmware.api [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for the task: (returnval){ [ 1079.185467] env[62952]: value = "task-1263837" [ 1079.185467] env[62952]: _type = "Task" [ 1079.185467] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.192419] env[62952]: DEBUG oslo_vmware.api [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263836, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.196916] env[62952]: DEBUG oslo_vmware.api [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263837, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.424889] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263830, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.549311} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.425206] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] f28d0043-236e-4df4-8478-e4b1befb6a7d/f28d0043-236e-4df4-8478-e4b1befb6a7d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1079.425430] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1079.425686] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a934c150-a82e-407d-b75b-57dd5a772d8d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.433197] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1079.433197] env[62952]: value = "task-1263838" [ 1079.433197] env[62952]: _type = "Task" [ 1079.433197] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.441870] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263838, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.639305] env[62952]: DEBUG nova.compute.manager [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1079.651849] env[62952]: DEBUG nova.network.neutron [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1079.665526] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263833, 'name': PowerOffVM_Task, 'duration_secs': 0.555831} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.665818] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1079.666610] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3834bfd9-8dc3-4233-af34-b0fd881cad0a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.700955] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-083a6b3a-3ce4-4a50-afef-4c7fadfc49c8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.704358] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e82779c1-7dd8-48bf-819c-cfbf1f646bc7 tempest-ServerAddressesTestJSON-1463833279 tempest-ServerAddressesTestJSON-1463833279-project-member] Lock "1a02fba4-aad0-4bd4-8183-a7f08d887a5a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.536s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.718668] env[62952]: DEBUG oslo_vmware.api [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Task: {'id': task-1263837, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.371008} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.718940] env[62952]: DEBUG oslo_vmware.api [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263836, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.37734} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.719455] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1079.719679] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1079.719851] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1079.720305] env[62952]: INFO nova.compute.manager [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1079.720305] env[62952]: DEBUG oslo.service.loopingcall [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1079.720508] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1079.720654] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1079.720827] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1079.721191] env[62952]: INFO nova.compute.manager [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1079.721228] env[62952]: DEBUG oslo.service.loopingcall [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1079.721655] env[62952]: DEBUG nova.compute.manager [-] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1079.721748] env[62952]: DEBUG nova.network.neutron [-] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1079.723482] env[62952]: DEBUG nova.compute.manager [-] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1079.723587] env[62952]: DEBUG nova.network.neutron [-] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1079.744785] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1079.745688] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-15760576-c922-43d6-9442-f3057d79f8f1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.751888] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1079.751888] env[62952]: value = "task-1263839" [ 1079.751888] env[62952]: _type = "Task" [ 1079.751888] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.760294] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263839, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.889872] env[62952]: DEBUG nova.network.neutron [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Updating instance_info_cache with network_info: [{"id": "116c22bf-b489-4039-9b75-598fe7be83dc", "address": "fa:16:3e:ef:50:20", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap116c22bf-b4", "ovs_interfaceid": "116c22bf-b489-4039-9b75-598fe7be83dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.935029] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.935342] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.935502] env[62952]: DEBUG nova.compute.manager [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Going to confirm migration 3 {{(pid=62952) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1079.946972] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263838, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081962} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.946972] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1079.947527] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f758bc-9d61-4699-86a6-fa916b610f83 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.970506] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] f28d0043-236e-4df4-8478-e4b1befb6a7d/f28d0043-236e-4df4-8478-e4b1befb6a7d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1079.975018] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f145bb2-d4a6-4e14-b91e-baded396abd4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.992991] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1079.992991] env[62952]: value = "task-1263840" [ 1079.992991] env[62952]: _type = "Task" [ 1079.992991] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.001563] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263840, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.126195] env[62952]: DEBUG nova.compute.manager [req-51f3054f-4319-4530-9865-a1f1e6090dc4 req-e29f860f-9f18-4b70-9934-af2b0213116d service nova] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Received event network-vif-deleted-cff4c81c-513a-42a4-b6f5-94b4367fa0ea {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1080.126320] env[62952]: INFO nova.compute.manager [req-51f3054f-4319-4530-9865-a1f1e6090dc4 req-e29f860f-9f18-4b70-9934-af2b0213116d service nova] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Neutron deleted interface cff4c81c-513a-42a4-b6f5-94b4367fa0ea; detaching it from the instance and deleting it from the info cache [ 1080.126408] env[62952]: DEBUG nova.network.neutron [req-51f3054f-4319-4530-9865-a1f1e6090dc4 req-e29f860f-9f18-4b70-9934-af2b0213116d service nova] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.165483] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.165833] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.167416] env[62952]: INFO nova.compute.claims [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1080.262325] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1080.262544] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1080.262799] env[62952]: DEBUG oslo_concurrency.lockutils [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.262971] env[62952]: DEBUG oslo_concurrency.lockutils [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.263170] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1080.263415] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-54797ca6-3c6d-44b0-a7cc-f6f37cdea974 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.272147] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1080.272364] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1080.273117] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f774ee2f-e10b-4d38-80f6-8a6e40b4cc05 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.279314] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1080.279314] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52902cca-2cf9-d12c-9784-e64b4afa04ee" [ 1080.279314] env[62952]: _type = "Task" [ 1080.279314] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.287565] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52902cca-2cf9-d12c-9784-e64b4afa04ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.394041] env[62952]: DEBUG oslo_concurrency.lockutils [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "refresh_cache-8cda6eea-d488-4d73-a7ab-7a905037929d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.394041] env[62952]: DEBUG nova.compute.manager [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Instance network_info: |[{"id": "116c22bf-b489-4039-9b75-598fe7be83dc", "address": "fa:16:3e:ef:50:20", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap116c22bf-b4", "ovs_interfaceid": "116c22bf-b489-4039-9b75-598fe7be83dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1080.394041] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:50:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '116c22bf-b489-4039-9b75-598fe7be83dc', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1080.402150] env[62952]: DEBUG oslo.service.loopingcall [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1080.402526] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1080.402828] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e9c3c7e8-c7e1-4be3-a60b-3921e3c60a90 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.426532] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1080.426532] env[62952]: value = "task-1263841" [ 1080.426532] env[62952]: _type = "Task" [ 1080.426532] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.435632] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263841, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.469592] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.470107] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.470319] env[62952]: DEBUG nova.network.neutron [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1080.470518] env[62952]: DEBUG nova.objects.instance [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lazy-loading 'info_cache' on Instance uuid 6e289c20-8dd2-4680-b816-9bbe82fe7aec {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1080.503118] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263840, 'name': ReconfigVM_Task, 'duration_secs': 0.425705} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.504789] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Reconfigured VM instance instance-00000063 to attach disk [datastore2] f28d0043-236e-4df4-8478-e4b1befb6a7d/f28d0043-236e-4df4-8478-e4b1befb6a7d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1080.505874] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa87013d-c5b8-4743-8288-b03ec11da3ee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.510869] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1080.510869] env[62952]: value = "task-1263842" [ 1080.510869] env[62952]: _type = "Task" [ 1080.510869] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.519804] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263842, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.527052] env[62952]: DEBUG nova.compute.manager [req-fcb49881-c157-4f9b-9dd1-e297c05f66f1 req-3d978858-16b0-4526-8c30-42466c5f0aa9 service nova] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Received event network-vif-deleted-a45070ed-ba02-43f8-b335-7cb961a6ad6d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1080.528337] env[62952]: INFO nova.compute.manager [req-fcb49881-c157-4f9b-9dd1-e297c05f66f1 req-3d978858-16b0-4526-8c30-42466c5f0aa9 service nova] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Neutron deleted interface a45070ed-ba02-43f8-b335-7cb961a6ad6d; detaching it from the instance and deleting it from the info cache [ 1080.528337] env[62952]: DEBUG nova.network.neutron [req-fcb49881-c157-4f9b-9dd1-e297c05f66f1 req-3d978858-16b0-4526-8c30-42466c5f0aa9 service nova] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.595809] env[62952]: DEBUG nova.network.neutron [-] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.632020] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9203ddd9-d27b-4fe7-b1cb-97fd268b3bb0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.639139] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5352b97d-664e-4cdb-a1ae-5249322653df {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.670026] env[62952]: DEBUG nova.compute.manager [req-51f3054f-4319-4530-9865-a1f1e6090dc4 req-e29f860f-9f18-4b70-9934-af2b0213116d service nova] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Detach interface failed, port_id=cff4c81c-513a-42a4-b6f5-94b4367fa0ea, reason: Instance e9e690b8-2fae-401d-aaad-0acef6e88ce3 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1080.790584] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52902cca-2cf9-d12c-9784-e64b4afa04ee, 'name': SearchDatastore_Task, 'duration_secs': 0.014806} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.791410] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc9fac95-97eb-4dca-9606-4bbfbd02702b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.797392] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1080.797392] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52089682-20c3-501f-8cb4-321196f36138" [ 1080.797392] env[62952]: _type = "Task" [ 1080.797392] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.805505] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52089682-20c3-501f-8cb4-321196f36138, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.936696] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263841, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.005927] env[62952]: DEBUG nova.network.neutron [-] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.027155] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263842, 'name': Rename_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.029945] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e6930e6-9939-4145-9836-e742fdc91dc6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.042192] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-225eacf4-dff7-468e-83d3-d8d36368f5b7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.072380] env[62952]: DEBUG nova.compute.manager [req-fcb49881-c157-4f9b-9dd1-e297c05f66f1 req-3d978858-16b0-4526-8c30-42466c5f0aa9 service nova] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Detach interface failed, port_id=a45070ed-ba02-43f8-b335-7cb961a6ad6d, reason: Instance e44914cd-d8ae-4f43-9207-32cacf6d6553 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1081.098551] env[62952]: INFO nova.compute.manager [-] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Took 1.37 seconds to deallocate network for instance. [ 1081.311929] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52089682-20c3-501f-8cb4-321196f36138, 'name': SearchDatastore_Task, 'duration_secs': 0.036605} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.311929] env[62952]: DEBUG oslo_concurrency.lockutils [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.311929] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 1941365e-d562-4d99-89eb-226a07e52071/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk. {{(pid=62952) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1081.312106] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-84fc890e-5ef0-410d-a061-38e1947282ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.318643] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1081.318643] env[62952]: value = "task-1263843" [ 1081.318643] env[62952]: _type = "Task" [ 1081.318643] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.326228] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263843, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.358447] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaeef66b-9fac-49b8-8d7b-53337251c86c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.366032] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b77022a-1c21-49c9-932d-427b361b57f7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.400915] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc93d69-19b5-4009-841d-c6de9a15ff3d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.410019] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30181b13-7aaf-4ac2-b17e-d45d199d7327 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.423234] env[62952]: DEBUG nova.compute.provider_tree [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.436116] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263841, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.510153] env[62952]: INFO nova.compute.manager [-] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Took 1.79 seconds to deallocate network for instance. [ 1081.524285] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263842, 'name': Rename_Task, 'duration_secs': 0.971318} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.524830] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1081.525578] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-67260834-66bb-4d0c-a980-2787c6c4e103 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.532667] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1081.532667] env[62952]: value = "task-1263844" [ 1081.532667] env[62952]: _type = "Task" [ 1081.532667] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.542675] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263844, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.605462] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.799731] env[62952]: DEBUG nova.network.neutron [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance_info_cache with network_info: [{"id": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "address": "fa:16:3e:eb:c2:8d", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e2e9d1e-dc", "ovs_interfaceid": "5e2e9d1e-dce4-404e-bf41-1a7897be3758", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.828884] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263843, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.927168] env[62952]: DEBUG nova.scheduler.client.report [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1081.940158] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263841, 'name': CreateVM_Task, 'duration_secs': 1.229697} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.940515] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1081.941040] env[62952]: DEBUG oslo_concurrency.lockutils [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.941225] env[62952]: DEBUG oslo_concurrency.lockutils [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.941570] env[62952]: DEBUG oslo_concurrency.lockutils [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1081.941834] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-683bfbf1-3a50-49cf-9fde-dec449e4f76a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.946822] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1081.946822] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]529b2f50-18be-eefb-ef63-4eea50192825" [ 1081.946822] env[62952]: _type = "Task" [ 1081.946822] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.955202] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]529b2f50-18be-eefb-ef63-4eea50192825, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.020115] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.042714] env[62952]: DEBUG oslo_vmware.api [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263844, 'name': PowerOnVM_Task, 'duration_secs': 0.47474} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.042983] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1082.043218] env[62952]: INFO nova.compute.manager [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Took 7.22 seconds to spawn the instance on the hypervisor. [ 1082.043401] env[62952]: DEBUG nova.compute.manager [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1082.044215] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19844ba-8356-4721-8b6b-6181913be7d1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.301652] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "refresh_cache-6e289c20-8dd2-4680-b816-9bbe82fe7aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.301950] env[62952]: DEBUG nova.objects.instance [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lazy-loading 'migration_context' on Instance uuid 6e289c20-8dd2-4680-b816-9bbe82fe7aec {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1082.330281] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263843, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517757} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.330571] env[62952]: INFO nova.virt.vmwareapi.ds_util [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 1941365e-d562-4d99-89eb-226a07e52071/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk. [ 1082.331432] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52628ff-8067-40b3-8748-715ab1ebfd8c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.360576] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 1941365e-d562-4d99-89eb-226a07e52071/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1082.360576] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84b29e2b-27bf-4e8f-9f38-3c1c6da0e4ea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.378028] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1082.378028] env[62952]: value = "task-1263845" [ 1082.378028] env[62952]: _type = "Task" [ 1082.378028] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.385689] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263845, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.432827] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.267s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.433385] env[62952]: DEBUG nova.compute.manager [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1082.436051] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.831s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.436251] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.438339] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.419s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.438554] env[62952]: DEBUG nova.objects.instance [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lazy-loading 'resources' on Instance uuid e44914cd-d8ae-4f43-9207-32cacf6d6553 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1082.457458] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]529b2f50-18be-eefb-ef63-4eea50192825, 'name': SearchDatastore_Task, 'duration_secs': 0.032302} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.457781] env[62952]: DEBUG oslo_concurrency.lockutils [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.457990] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1082.458238] env[62952]: DEBUG oslo_concurrency.lockutils [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.458388] env[62952]: DEBUG oslo_concurrency.lockutils [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.458565] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1082.458869] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6d66bd79-a663-41e3-905f-a5250059bc13 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.461351] env[62952]: INFO nova.scheduler.client.report [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleted allocations for instance e9e690b8-2fae-401d-aaad-0acef6e88ce3 [ 1082.474131] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1082.474311] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1082.475230] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-768ebc07-3e43-4bc9-b771-fb1cf22fd44b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.480122] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1082.480122] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52183e91-d556-e115-b84f-b25118267c21" [ 1082.480122] env[62952]: _type = "Task" [ 1082.480122] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.487318] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52183e91-d556-e115-b84f-b25118267c21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.558879] env[62952]: INFO nova.compute.manager [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Took 13.02 seconds to build instance. [ 1082.804764] env[62952]: DEBUG nova.objects.base [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Object Instance<6e289c20-8dd2-4680-b816-9bbe82fe7aec> lazy-loaded attributes: info_cache,migration_context {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1082.805784] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6fdf7b-eeb6-4245-aba6-5798b7d8e6f5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.824580] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b053b1be-6481-4b01-95f9-034c7673d7a2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.829987] env[62952]: DEBUG oslo_vmware.api [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1082.829987] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526f6ab3-fb65-aba0-2a7c-f85f9a9caa6a" [ 1082.829987] env[62952]: _type = "Task" [ 1082.829987] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.837272] env[62952]: DEBUG oslo_vmware.api [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526f6ab3-fb65-aba0-2a7c-f85f9a9caa6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.887097] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263845, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.942039] env[62952]: DEBUG nova.compute.utils [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1082.946195] env[62952]: DEBUG nova.compute.manager [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1082.946873] env[62952]: DEBUG nova.network.neutron [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1082.968099] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e1c09ac-e02e-45e6-86fe-277605135a49 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "e9e690b8-2fae-401d-aaad-0acef6e88ce3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.435s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.991792] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52183e91-d556-e115-b84f-b25118267c21, 'name': SearchDatastore_Task, 'duration_secs': 0.021398} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.993175] env[62952]: DEBUG nova.policy [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3ec1aecc0cf64aaea875a4280b114874', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fbb5ccb4d615429facd6932fb55b04e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1082.995060] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c3c5a67-de8e-4f67-8626-6c1a19bec033 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.000988] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1083.000988] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ce39dd-1fdb-7d55-b9d6-55e968c5591e" [ 1083.000988] env[62952]: _type = "Task" [ 1083.000988] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.010915] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ce39dd-1fdb-7d55-b9d6-55e968c5591e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.061883] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b09ef063-e848-475a-b57a-8cf8b8597702 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "f28d0043-236e-4df4-8478-e4b1befb6a7d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.536s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.099471] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ccebe8-eeba-4c27-ac35-854234e66f43 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.106765] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293c2b87-9255-416e-bc1d-7d64865c9693 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.137965] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1064faf-a7f7-493f-b4bd-821b6ee68cc1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.146048] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a53b7e36-7051-4040-9906-92d629ec866c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.159153] env[62952]: DEBUG nova.compute.provider_tree [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.257193] env[62952]: DEBUG nova.network.neutron [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Successfully created port: 7f9fb655-8880-45a0-9c96-75a3faf48d12 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1083.343335] env[62952]: DEBUG oslo_vmware.api [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]526f6ab3-fb65-aba0-2a7c-f85f9a9caa6a, 'name': SearchDatastore_Task, 'duration_secs': 0.015185} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.344051] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.392065] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263845, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.446583] env[62952]: DEBUG nova.compute.manager [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1083.514643] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ce39dd-1fdb-7d55-b9d6-55e968c5591e, 'name': SearchDatastore_Task, 'duration_secs': 0.020179} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.514925] env[62952]: DEBUG oslo_concurrency.lockutils [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.515199] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 8cda6eea-d488-4d73-a7ab-7a905037929d/8cda6eea-d488-4d73-a7ab-7a905037929d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1083.515467] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d4010d4-e872-498c-9b7d-9966993f0d54 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.523836] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1083.523836] env[62952]: value = "task-1263846" [ 1083.523836] env[62952]: _type = "Task" [ 1083.523836] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.532575] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263846, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.662368] env[62952]: DEBUG nova.scheduler.client.report [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1083.889378] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263845, 'name': ReconfigVM_Task, 'duration_secs': 1.144893} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.890159] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 1941365e-d562-4d99-89eb-226a07e52071/119dd2b4-b6de-42d5-adba-db818e4c2cd7-rescue.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1083.891346] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b874231-533c-41aa-bf27-388611ea755b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.920089] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-020681fc-0be7-46e1-826b-f1c1ab001d98 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.936063] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1083.936063] env[62952]: value = "task-1263847" [ 1083.936063] env[62952]: _type = "Task" [ 1083.936063] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.944690] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263847, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.985508] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e751291c-63a3-4009-9333-98e323192f74 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "f28d0043-236e-4df4-8478-e4b1befb6a7d" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.986201] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e751291c-63a3-4009-9333-98e323192f74 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "f28d0043-236e-4df4-8478-e4b1befb6a7d" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.986589] env[62952]: DEBUG nova.compute.manager [None req-e751291c-63a3-4009-9333-98e323192f74 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1083.988179] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d36893-302d-4b27-9d70-1e50e3c88320 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.997471] env[62952]: DEBUG nova.compute.manager [None req-e751291c-63a3-4009-9333-98e323192f74 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62952) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1083.998467] env[62952]: DEBUG nova.objects.instance [None req-e751291c-63a3-4009-9333-98e323192f74 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lazy-loading 'flavor' on Instance uuid f28d0043-236e-4df4-8478-e4b1befb6a7d {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1084.037543] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263846, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.168063] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.729s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.170164] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.826s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.189022] env[62952]: INFO nova.scheduler.client.report [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Deleted allocations for instance e44914cd-d8ae-4f43-9207-32cacf6d6553 [ 1084.448246] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263847, 'name': ReconfigVM_Task, 'duration_secs': 0.325162} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.448597] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1084.448882] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe10da1e-2fb1-489f-9e57-bafed65ee257 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.455232] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1084.455232] env[62952]: value = "task-1263848" [ 1084.455232] env[62952]: _type = "Task" [ 1084.455232] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.459106] env[62952]: DEBUG nova.compute.manager [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1084.465386] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263848, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.483828] env[62952]: DEBUG nova.virt.hardware [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1084.484104] env[62952]: DEBUG nova.virt.hardware [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1084.484271] env[62952]: DEBUG nova.virt.hardware [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1084.484458] env[62952]: DEBUG nova.virt.hardware [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1084.484609] env[62952]: DEBUG nova.virt.hardware [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1084.484812] env[62952]: DEBUG nova.virt.hardware [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1084.485101] env[62952]: DEBUG nova.virt.hardware [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1084.485283] env[62952]: DEBUG nova.virt.hardware [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1084.485456] env[62952]: DEBUG nova.virt.hardware [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1084.485636] env[62952]: DEBUG nova.virt.hardware [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1084.485819] env[62952]: DEBUG nova.virt.hardware [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1084.486766] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5de788d-4869-48e9-9a33-8c8b86b7da63 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.494511] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa6ac3a-de69-480c-8112-5bacb0210f71 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.512395] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e751291c-63a3-4009-9333-98e323192f74 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1084.512691] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-afd39013-e440-48da-a21e-96c31e039857 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.519339] env[62952]: DEBUG oslo_vmware.api [None req-e751291c-63a3-4009-9333-98e323192f74 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1084.519339] env[62952]: value = "task-1263849" [ 1084.519339] env[62952]: _type = "Task" [ 1084.519339] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.528744] env[62952]: DEBUG oslo_vmware.api [None req-e751291c-63a3-4009-9333-98e323192f74 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263849, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.537104] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263846, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.621605} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.537410] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 8cda6eea-d488-4d73-a7ab-7a905037929d/8cda6eea-d488-4d73-a7ab-7a905037929d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1084.537633] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1084.537929] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5a28b9eb-d4f9-4d20-a6f1-55e4fc0afc74 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.544045] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1084.544045] env[62952]: value = "task-1263850" [ 1084.544045] env[62952]: _type = "Task" [ 1084.544045] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.551626] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263850, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.687969] env[62952]: DEBUG nova.compute.manager [req-1b68802b-afac-4ad4-bc52-143609c82bc1 req-4503d09e-b794-42f5-9b8b-43398dbe6825 service nova] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Received event network-vif-plugged-7f9fb655-8880-45a0-9c96-75a3faf48d12 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1084.688236] env[62952]: DEBUG oslo_concurrency.lockutils [req-1b68802b-afac-4ad4-bc52-143609c82bc1 req-4503d09e-b794-42f5-9b8b-43398dbe6825 service nova] Acquiring lock "126f57b5-be03-4164-b92f-3c2e96dd9171-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.688436] env[62952]: DEBUG oslo_concurrency.lockutils [req-1b68802b-afac-4ad4-bc52-143609c82bc1 req-4503d09e-b794-42f5-9b8b-43398dbe6825 service nova] Lock "126f57b5-be03-4164-b92f-3c2e96dd9171-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.688601] env[62952]: DEBUG oslo_concurrency.lockutils [req-1b68802b-afac-4ad4-bc52-143609c82bc1 req-4503d09e-b794-42f5-9b8b-43398dbe6825 service nova] Lock "126f57b5-be03-4164-b92f-3c2e96dd9171-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.688823] env[62952]: DEBUG nova.compute.manager [req-1b68802b-afac-4ad4-bc52-143609c82bc1 req-4503d09e-b794-42f5-9b8b-43398dbe6825 service nova] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] No waiting events found dispatching network-vif-plugged-7f9fb655-8880-45a0-9c96-75a3faf48d12 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1084.688947] env[62952]: WARNING nova.compute.manager [req-1b68802b-afac-4ad4-bc52-143609c82bc1 req-4503d09e-b794-42f5-9b8b-43398dbe6825 service nova] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Received unexpected event network-vif-plugged-7f9fb655-8880-45a0-9c96-75a3faf48d12 for instance with vm_state building and task_state spawning. [ 1084.698092] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51463949-efbf-4b09-a225-1d2aeefd2060 tempest-AttachVolumeTestJSON-1448152293 tempest-AttachVolumeTestJSON-1448152293-project-member] Lock "e44914cd-d8ae-4f43-9207-32cacf6d6553" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.148s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.812507] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9545a0c6-861a-4465-bd5c-49f0b8a9d871 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.820736] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f981b96c-06a8-4ea4-b833-c15aac80bc13 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.825386] env[62952]: DEBUG nova.network.neutron [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Successfully updated port: 7f9fb655-8880-45a0-9c96-75a3faf48d12 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1084.857859] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29695f82-da06-4a12-8c4b-f8c17d4a2aac {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.866334] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb7497b6-c6c3-4241-9c65-2132042da9fd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.880881] env[62952]: DEBUG nova.compute.provider_tree [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.965172] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263848, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.029855] env[62952]: DEBUG oslo_vmware.api [None req-e751291c-63a3-4009-9333-98e323192f74 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263849, 'name': PowerOffVM_Task, 'duration_secs': 0.19596} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.029855] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e751291c-63a3-4009-9333-98e323192f74 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1085.030081] env[62952]: DEBUG nova.compute.manager [None req-e751291c-63a3-4009-9333-98e323192f74 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1085.030756] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82795c63-6daf-4e6d-9814-d8fc78238247 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.052601] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263850, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084701} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.052811] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1085.053568] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d228287-6c6a-484b-870b-ea57a062ef56 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.077838] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 8cda6eea-d488-4d73-a7ab-7a905037929d/8cda6eea-d488-4d73-a7ab-7a905037929d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1085.077971] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a6e67dc-271b-4dab-8fd5-565ded2755fe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.098256] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1085.098256] env[62952]: value = "task-1263851" [ 1085.098256] env[62952]: _type = "Task" [ 1085.098256] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.106558] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263851, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.328962] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "refresh_cache-126f57b5-be03-4164-b92f-3c2e96dd9171" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.329197] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "refresh_cache-126f57b5-be03-4164-b92f-3c2e96dd9171" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.329336] env[62952]: DEBUG nova.network.neutron [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1085.383875] env[62952]: DEBUG nova.scheduler.client.report [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1085.467231] env[62952]: DEBUG oslo_vmware.api [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263848, 'name': PowerOnVM_Task, 'duration_secs': 0.526204} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.467517] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1085.470308] env[62952]: DEBUG nova.compute.manager [None req-85cb4e0c-18f2-4d1d-a021-561bc2d42222 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1085.471128] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b35f9e13-e934-433c-aeb4-2dca61ebaca9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.541987] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e751291c-63a3-4009-9333-98e323192f74 tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "f28d0043-236e-4df4-8478-e4b1befb6a7d" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.556s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.610029] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263851, 'name': ReconfigVM_Task, 'duration_secs': 0.286663} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.610336] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 8cda6eea-d488-4d73-a7ab-7a905037929d/8cda6eea-d488-4d73-a7ab-7a905037929d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1085.610940] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-49511014-a69f-4e16-9bde-96cd4355d39c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.617588] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1085.617588] env[62952]: value = "task-1263853" [ 1085.617588] env[62952]: _type = "Task" [ 1085.617588] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.625781] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263853, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.860717] env[62952]: DEBUG nova.network.neutron [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1086.008585] env[62952]: DEBUG nova.network.neutron [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Updating instance_info_cache with network_info: [{"id": "7f9fb655-8880-45a0-9c96-75a3faf48d12", "address": "fa:16:3e:41:35:a3", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f9fb655-88", "ovs_interfaceid": "7f9fb655-8880-45a0-9c96-75a3faf48d12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.132301] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263853, 'name': Rename_Task, 'duration_secs': 0.135834} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.132803] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1086.133203] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aa92750e-db77-4232-9beb-6a812f3fcb02 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.140161] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1086.140161] env[62952]: value = "task-1263854" [ 1086.140161] env[62952]: _type = "Task" [ 1086.140161] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.148316] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263854, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.289609] env[62952]: INFO nova.compute.manager [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Unrescuing [ 1086.289890] env[62952]: DEBUG oslo_concurrency.lockutils [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.290374] env[62952]: DEBUG oslo_concurrency.lockutils [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquired lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.290575] env[62952]: DEBUG nova.network.neutron [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1086.396903] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.226s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.510964] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "refresh_cache-126f57b5-be03-4164-b92f-3c2e96dd9171" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.511342] env[62952]: DEBUG nova.compute.manager [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Instance network_info: |[{"id": "7f9fb655-8880-45a0-9c96-75a3faf48d12", "address": "fa:16:3e:41:35:a3", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f9fb655-88", "ovs_interfaceid": "7f9fb655-8880-45a0-9c96-75a3faf48d12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1086.511792] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:41:35:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5ad8894f-e240-4013-8272-4e79daea0751', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7f9fb655-8880-45a0-9c96-75a3faf48d12', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1086.519815] env[62952]: DEBUG oslo.service.loopingcall [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1086.520075] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1086.520317] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-362e752f-14fa-4f61-90ff-e9a2bb0bf922 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.540134] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1086.540134] env[62952]: value = "task-1263855" [ 1086.540134] env[62952]: _type = "Task" [ 1086.540134] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.548771] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263855, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.652498] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263854, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.714453] env[62952]: DEBUG nova.compute.manager [req-5289bf0c-602a-414a-8f60-dfccbb4456c0 req-03988d8c-dbf0-4267-9f62-9b0a0aca39ce service nova] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Received event network-changed-7f9fb655-8880-45a0-9c96-75a3faf48d12 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1086.714625] env[62952]: DEBUG nova.compute.manager [req-5289bf0c-602a-414a-8f60-dfccbb4456c0 req-03988d8c-dbf0-4267-9f62-9b0a0aca39ce service nova] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Refreshing instance network info cache due to event network-changed-7f9fb655-8880-45a0-9c96-75a3faf48d12. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1086.714808] env[62952]: DEBUG oslo_concurrency.lockutils [req-5289bf0c-602a-414a-8f60-dfccbb4456c0 req-03988d8c-dbf0-4267-9f62-9b0a0aca39ce service nova] Acquiring lock "refresh_cache-126f57b5-be03-4164-b92f-3c2e96dd9171" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.714959] env[62952]: DEBUG oslo_concurrency.lockutils [req-5289bf0c-602a-414a-8f60-dfccbb4456c0 req-03988d8c-dbf0-4267-9f62-9b0a0aca39ce service nova] Acquired lock "refresh_cache-126f57b5-be03-4164-b92f-3c2e96dd9171" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.715168] env[62952]: DEBUG nova.network.neutron [req-5289bf0c-602a-414a-8f60-dfccbb4456c0 req-03988d8c-dbf0-4267-9f62-9b0a0aca39ce service nova] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Refreshing network info cache for port 7f9fb655-8880-45a0-9c96-75a3faf48d12 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1086.717350] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "f28d0043-236e-4df4-8478-e4b1befb6a7d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.717350] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "f28d0043-236e-4df4-8478-e4b1befb6a7d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.717480] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "f28d0043-236e-4df4-8478-e4b1befb6a7d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.717655] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "f28d0043-236e-4df4-8478-e4b1befb6a7d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.717867] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "f28d0043-236e-4df4-8478-e4b1befb6a7d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.720006] env[62952]: INFO nova.compute.manager [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Terminating instance [ 1086.721778] env[62952]: DEBUG nova.compute.manager [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1086.721980] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1086.722848] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d19562-ab0b-4d16-a7b3-1efb2e5b6906 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.733432] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1086.733778] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1a9537aa-d4e6-4d05-8a98-38799fa6bae3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.813397] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1086.813553] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1086.813772] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleting the datastore file [datastore2] f28d0043-236e-4df4-8478-e4b1befb6a7d {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1086.814069] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1ca9361e-43c0-465b-ae18-132fc3379f0b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.824660] env[62952]: DEBUG oslo_vmware.api [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for the task: (returnval){ [ 1086.824660] env[62952]: value = "task-1263857" [ 1086.824660] env[62952]: _type = "Task" [ 1086.824660] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.837795] env[62952]: DEBUG oslo_vmware.api [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263857, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.976375] env[62952]: INFO nova.scheduler.client.report [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Deleted allocation for migration 8ba8486b-d3a9-43a5-9640-d6bc578ae322 [ 1087.055772] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263855, 'name': CreateVM_Task, 'duration_secs': 0.356452} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.056402] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1087.056739] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.056881] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.057221] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1087.057489] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3deabd0-acf6-4f5c-8ecc-6d7d6f0043fe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.062898] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1087.062898] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e1bd63-86c6-4a02-733a-a14cc51cee53" [ 1087.062898] env[62952]: _type = "Task" [ 1087.062898] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.072430] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e1bd63-86c6-4a02-733a-a14cc51cee53, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.088014] env[62952]: DEBUG nova.network.neutron [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Updating instance_info_cache with network_info: [{"id": "e1c235af-9501-47c0-a83a-ba40ea7fa8c1", "address": "fa:16:3e:34:07:9b", "network": {"id": "827a8171-6759-4434-a6f1-437aad881c81", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1902486970-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091cdeb6da48492bb02f93822a45c9bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11b669be-fb26-4ef8-bdb6-c77ab9d06daf", "external-id": "nsx-vlan-transportzone-633", "segmentation_id": 633, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1c235af-95", "ovs_interfaceid": "e1c235af-9501-47c0-a83a-ba40ea7fa8c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.152367] env[62952]: DEBUG oslo_vmware.api [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263854, 'name': PowerOnVM_Task, 'duration_secs': 0.820496} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.152680] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1087.152910] env[62952]: INFO nova.compute.manager [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Took 10.03 seconds to spawn the instance on the hypervisor. [ 1087.153119] env[62952]: DEBUG nova.compute.manager [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1087.153991] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8527746-32bf-4cd9-9c2f-45013b34b441 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.333985] env[62952]: DEBUG oslo_vmware.api [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Task: {'id': task-1263857, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.277259} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.334272] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1087.334480] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1087.334785] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1087.334855] env[62952]: INFO nova.compute.manager [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1087.335094] env[62952]: DEBUG oslo.service.loopingcall [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1087.335301] env[62952]: DEBUG nova.compute.manager [-] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1087.335400] env[62952]: DEBUG nova.network.neutron [-] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1087.437697] env[62952]: DEBUG nova.network.neutron [req-5289bf0c-602a-414a-8f60-dfccbb4456c0 req-03988d8c-dbf0-4267-9f62-9b0a0aca39ce service nova] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Updated VIF entry in instance network info cache for port 7f9fb655-8880-45a0-9c96-75a3faf48d12. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1087.438091] env[62952]: DEBUG nova.network.neutron [req-5289bf0c-602a-414a-8f60-dfccbb4456c0 req-03988d8c-dbf0-4267-9f62-9b0a0aca39ce service nova] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Updating instance_info_cache with network_info: [{"id": "7f9fb655-8880-45a0-9c96-75a3faf48d12", "address": "fa:16:3e:41:35:a3", "network": {"id": "16eb8ebc-94bb-4ed0-9e8c-19926c0197d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-711970400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fbb5ccb4d615429facd6932fb55b04e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ad8894f-e240-4013-8272-4e79daea0751", "external-id": "nsx-vlan-transportzone-204", "segmentation_id": 204, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f9fb655-88", "ovs_interfaceid": "7f9fb655-8880-45a0-9c96-75a3faf48d12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.446221] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.485271] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6fa9de1-6359-48d2-ac94-65443c661810 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.550s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.486880] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.041s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.487130] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.487339] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.487539] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.490880] env[62952]: INFO nova.compute.manager [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Terminating instance [ 1087.492729] env[62952]: DEBUG nova.compute.manager [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1087.492933] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1087.493834] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d0fa540-b365-48a4-9683-b53b9424e7e7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.501494] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1087.502261] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c893672e-1d1b-455a-8839-811aa9e99098 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.508483] env[62952]: DEBUG oslo_vmware.api [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1087.508483] env[62952]: value = "task-1263858" [ 1087.508483] env[62952]: _type = "Task" [ 1087.508483] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.517558] env[62952]: DEBUG oslo_vmware.api [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263858, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.576326] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e1bd63-86c6-4a02-733a-a14cc51cee53, 'name': SearchDatastore_Task, 'duration_secs': 0.01013} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.576800] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.576906] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1087.577185] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.577332] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.577518] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1087.577936] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2f49a2ad-3bfd-4661-9656-1f07a8515873 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.586920] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1087.587163] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1087.587968] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2933a3c4-0883-499c-a553-5ca2f6764d65 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.591423] env[62952]: DEBUG oslo_concurrency.lockutils [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Releasing lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.592228] env[62952]: DEBUG nova.objects.instance [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lazy-loading 'flavor' on Instance uuid 1941365e-d562-4d99-89eb-226a07e52071 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.597135] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1087.597135] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525bee02-02b9-ff9d-6423-29da994e6433" [ 1087.597135] env[62952]: _type = "Task" [ 1087.597135] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.608869] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525bee02-02b9-ff9d-6423-29da994e6433, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.623287] env[62952]: DEBUG nova.compute.manager [req-c8cf8416-d1c3-4e2f-b443-e8463bb371a1 req-0832b984-498a-4e7a-9061-e969bc6f46af service nova] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Received event network-vif-deleted-c764e89c-8999-4422-ab55-e64bf2dee40b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1087.623510] env[62952]: INFO nova.compute.manager [req-c8cf8416-d1c3-4e2f-b443-e8463bb371a1 req-0832b984-498a-4e7a-9061-e969bc6f46af service nova] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Neutron deleted interface c764e89c-8999-4422-ab55-e64bf2dee40b; detaching it from the instance and deleting it from the info cache [ 1087.623693] env[62952]: DEBUG nova.network.neutron [req-c8cf8416-d1c3-4e2f-b443-e8463bb371a1 req-0832b984-498a-4e7a-9061-e969bc6f46af service nova] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.674017] env[62952]: INFO nova.compute.manager [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Took 14.97 seconds to build instance. [ 1087.940879] env[62952]: DEBUG oslo_concurrency.lockutils [req-5289bf0c-602a-414a-8f60-dfccbb4456c0 req-03988d8c-dbf0-4267-9f62-9b0a0aca39ce service nova] Releasing lock "refresh_cache-126f57b5-be03-4164-b92f-3c2e96dd9171" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.020844] env[62952]: DEBUG oslo_vmware.api [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263858, 'name': PowerOffVM_Task, 'duration_secs': 0.289722} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.021241] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1088.021466] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1088.021817] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e7c25cde-b290-4206-93a7-6280381b2fa7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.098384] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6429f7c5-dba1-4113-a726-f7d4e0ad4fd2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.104074] env[62952]: DEBUG nova.network.neutron [-] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.107211] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1088.107537] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1088.107810] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Deleting the datastore file [datastore1] 6e289c20-8dd2-4680-b816-9bbe82fe7aec {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1088.127012] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6fcc40d6-b027-4ae8-9e7a-16e910d1a794 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.129558] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1088.133313] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eaad0988-47bf-432d-a65f-3c7b2988c5de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.135097] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-57dc956e-cf31-4f39-89e2-ee0157802629 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.136729] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525bee02-02b9-ff9d-6423-29da994e6433, 'name': SearchDatastore_Task, 'duration_secs': 0.011491} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.138751] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d574d9d-f81e-4a61-acbd-01a6a6aad214 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.142407] env[62952]: DEBUG oslo_vmware.api [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1088.142407] env[62952]: value = "task-1263860" [ 1088.142407] env[62952]: _type = "Task" [ 1088.142407] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.144645] env[62952]: DEBUG oslo_vmware.api [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1088.144645] env[62952]: value = "task-1263861" [ 1088.144645] env[62952]: _type = "Task" [ 1088.144645] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.150538] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cda0b61-db23-4c1f-8a7a-1bf13fd8fec1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.166089] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1088.166089] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52251ee8-8224-9aaa-39e6-42ec89142544" [ 1088.166089] env[62952]: _type = "Task" [ 1088.166089] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.179784] env[62952]: DEBUG oslo_concurrency.lockutils [None req-930bb38a-c658-4070-ace1-e7c0e9f0dd0f tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "8cda6eea-d488-4d73-a7ab-7a905037929d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.492s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.180064] env[62952]: DEBUG oslo_vmware.api [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263861, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.180286] env[62952]: DEBUG oslo_vmware.api [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263860, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.190499] env[62952]: DEBUG nova.compute.manager [req-c8cf8416-d1c3-4e2f-b443-e8463bb371a1 req-0832b984-498a-4e7a-9061-e969bc6f46af service nova] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Detach interface failed, port_id=c764e89c-8999-4422-ab55-e64bf2dee40b, reason: Instance f28d0043-236e-4df4-8478-e4b1befb6a7d could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1088.197634] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52251ee8-8224-9aaa-39e6-42ec89142544, 'name': SearchDatastore_Task, 'duration_secs': 0.023822} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.197984] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.198320] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 126f57b5-be03-4164-b92f-3c2e96dd9171/126f57b5-be03-4164-b92f-3c2e96dd9171.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1088.198661] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7c1b74e4-5341-440c-a3c0-8606693142cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.205854] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1088.205854] env[62952]: value = "task-1263862" [ 1088.205854] env[62952]: _type = "Task" [ 1088.205854] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.214151] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263862, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.606671] env[62952]: INFO nova.compute.manager [-] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Took 1.27 seconds to deallocate network for instance. [ 1088.657653] env[62952]: DEBUG oslo_vmware.api [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263860, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.195399} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.660830] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1088.661054] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1088.661246] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1088.661425] env[62952]: INFO nova.compute.manager [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1088.661677] env[62952]: DEBUG oslo.service.loopingcall [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1088.661901] env[62952]: DEBUG oslo_vmware.api [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263861, 'name': PowerOffVM_Task, 'duration_secs': 0.209047} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.662124] env[62952]: DEBUG nova.compute.manager [-] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1088.662227] env[62952]: DEBUG nova.network.neutron [-] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1088.664016] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1088.669335] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Reconfiguring VM instance instance-00000059 to detach disk 2002 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1088.670034] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33a07774-a947-49a9-9773-e7521be7bd68 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.689299] env[62952]: DEBUG oslo_vmware.api [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1088.689299] env[62952]: value = "task-1263864" [ 1088.689299] env[62952]: _type = "Task" [ 1088.689299] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.697850] env[62952]: DEBUG oslo_vmware.api [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263864, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.715192] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263862, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.917642] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed75033-2fab-4d24-95ec-a6034bacffea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.924551] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d8c732-2668-47ec-8302-ffcd923349e7 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Suspending the VM {{(pid=62952) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1088.924804] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-b3d43075-3056-4a61-b853-50dad9c581f5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.930993] env[62952]: DEBUG oslo_vmware.api [None req-f3d8c732-2668-47ec-8302-ffcd923349e7 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1088.930993] env[62952]: value = "task-1263865" [ 1088.930993] env[62952]: _type = "Task" [ 1088.930993] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.938874] env[62952]: DEBUG oslo_vmware.api [None req-f3d8c732-2668-47ec-8302-ffcd923349e7 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263865, 'name': SuspendVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.114051] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.114432] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.114583] env[62952]: DEBUG nova.objects.instance [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lazy-loading 'resources' on Instance uuid f28d0043-236e-4df4-8478-e4b1befb6a7d {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.199762] env[62952]: DEBUG oslo_vmware.api [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263864, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.216944] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263862, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522959} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.217293] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 126f57b5-be03-4164-b92f-3c2e96dd9171/126f57b5-be03-4164-b92f-3c2e96dd9171.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1089.217579] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1089.217825] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aad70cc5-b1ae-49d9-85d8-e5d6ddf6d16e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.224961] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1089.224961] env[62952]: value = "task-1263866" [ 1089.224961] env[62952]: _type = "Task" [ 1089.224961] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.235477] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263866, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.440957] env[62952]: DEBUG oslo_vmware.api [None req-f3d8c732-2668-47ec-8302-ffcd923349e7 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263865, 'name': SuspendVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.653024] env[62952]: DEBUG nova.compute.manager [req-93515e0c-3dcf-4d23-b6a9-36d96489fc45 req-d33e480f-13ae-4bbc-9979-d9046a921508 service nova] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Received event network-vif-deleted-5e2e9d1e-dce4-404e-bf41-1a7897be3758 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1089.653242] env[62952]: INFO nova.compute.manager [req-93515e0c-3dcf-4d23-b6a9-36d96489fc45 req-d33e480f-13ae-4bbc-9979-d9046a921508 service nova] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Neutron deleted interface 5e2e9d1e-dce4-404e-bf41-1a7897be3758; detaching it from the instance and deleting it from the info cache [ 1089.653420] env[62952]: DEBUG nova.network.neutron [req-93515e0c-3dcf-4d23-b6a9-36d96489fc45 req-d33e480f-13ae-4bbc-9979-d9046a921508 service nova] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.658732] env[62952]: DEBUG nova.network.neutron [-] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.701839] env[62952]: DEBUG oslo_vmware.api [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263864, 'name': ReconfigVM_Task, 'duration_secs': 0.609069} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.702163] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Reconfigured VM instance instance-00000059 to detach disk 2002 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1089.702366] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1089.702908] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac410821-fe8a-47f7-aa52-f9901000a011 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.708650] env[62952]: DEBUG oslo_vmware.api [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1089.708650] env[62952]: value = "task-1263867" [ 1089.708650] env[62952]: _type = "Task" [ 1089.708650] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.716324] env[62952]: DEBUG oslo_vmware.api [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263867, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.736709] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263866, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069862} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.736709] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1089.737312] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33bb5d78-30ae-468e-b719-ef5aa46c7035 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.741940] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ec2d55-c972-4896-8dc2-c7f50c45d407 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.764788] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 126f57b5-be03-4164-b92f-3c2e96dd9171/126f57b5-be03-4164-b92f-3c2e96dd9171.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1089.769951] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f7b518d-351e-4faf-87cf-e6bc0a60c740 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.785697] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa5a2e7c-2e55-4a80-b99d-89d43987282e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.822869] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1089.822869] env[62952]: value = "task-1263868" [ 1089.822869] env[62952]: _type = "Task" [ 1089.822869] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.823627] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c214976c-e2ae-49bd-8bac-e3ea34f1b581 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.835972] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263868, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.837243] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbcc747b-6ea9-4754-907d-5b24fa6125ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.851530] env[62952]: DEBUG nova.compute.provider_tree [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1089.941399] env[62952]: DEBUG oslo_vmware.api [None req-f3d8c732-2668-47ec-8302-ffcd923349e7 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263865, 'name': SuspendVM_Task, 'duration_secs': 0.64073} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.941633] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d8c732-2668-47ec-8302-ffcd923349e7 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Suspended the VM {{(pid=62952) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1089.941818] env[62952]: DEBUG nova.compute.manager [None req-f3d8c732-2668-47ec-8302-ffcd923349e7 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1089.942606] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eb3dbf9-a916-4e55-98a7-9ee5e93e1ff9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.158663] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dfedea5c-23cd-40f1-992c-39bf3a5b4c53 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.160958] env[62952]: INFO nova.compute.manager [-] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Took 1.50 seconds to deallocate network for instance. [ 1090.169796] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c66978-b897-46fa-9181-a524a72e6be9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.197685] env[62952]: DEBUG nova.compute.manager [req-93515e0c-3dcf-4d23-b6a9-36d96489fc45 req-d33e480f-13ae-4bbc-9979-d9046a921508 service nova] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Detach interface failed, port_id=5e2e9d1e-dce4-404e-bf41-1a7897be3758, reason: Instance 6e289c20-8dd2-4680-b816-9bbe82fe7aec could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1090.218184] env[62952]: DEBUG oslo_vmware.api [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263867, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.337167] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263868, 'name': ReconfigVM_Task, 'duration_secs': 0.260143} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.337665] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 126f57b5-be03-4164-b92f-3c2e96dd9171/126f57b5-be03-4164-b92f-3c2e96dd9171.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1090.338149] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2051520e-0816-41e8-b1ca-019d32e6411a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.345068] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1090.345068] env[62952]: value = "task-1263869" [ 1090.345068] env[62952]: _type = "Task" [ 1090.345068] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.355250] env[62952]: DEBUG nova.scheduler.client.report [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1090.358334] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263869, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.669093] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.719123] env[62952]: DEBUG oslo_vmware.api [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263867, 'name': PowerOnVM_Task, 'duration_secs': 0.898764} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.719424] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1090.719662] env[62952]: DEBUG nova.compute.manager [None req-981959df-e653-43b5-9dc3-5fa7c56e444f tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1090.720428] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69e64ca-b3fa-406b-809c-50ca870e4d0e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.855022] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263869, 'name': Rename_Task, 'duration_secs': 0.139921} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.855022] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1090.855022] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-28b2b346-1a4c-41d2-bc98-47c26e3e59ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.859363] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.745s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.862627] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.194s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.862867] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.864788] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1090.864788] env[62952]: value = "task-1263871" [ 1090.864788] env[62952]: _type = "Task" [ 1090.864788] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.873565] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263871, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.883236] env[62952]: INFO nova.scheduler.client.report [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Deleted allocations for instance 6e289c20-8dd2-4680-b816-9bbe82fe7aec [ 1090.885332] env[62952]: INFO nova.scheduler.client.report [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Deleted allocations for instance f28d0043-236e-4df4-8478-e4b1befb6a7d [ 1091.320572] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "8cda6eea-d488-4d73-a7ab-7a905037929d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.320859] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "8cda6eea-d488-4d73-a7ab-7a905037929d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.321225] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "8cda6eea-d488-4d73-a7ab-7a905037929d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.321436] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "8cda6eea-d488-4d73-a7ab-7a905037929d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.321618] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "8cda6eea-d488-4d73-a7ab-7a905037929d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.323802] env[62952]: INFO nova.compute.manager [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Terminating instance [ 1091.325717] env[62952]: DEBUG nova.compute.manager [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1091.325919] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1091.326757] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bee932d-83c7-4df8-aa07-ec3fa1394c76 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.334693] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1091.334928] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-37948f6d-9dfe-468b-abac-b87f789d236c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.375686] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263871, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.397311] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30d8de9d-474c-46be-87f5-288be6c74c0e tempest-ServersTestJSON-657684837 tempest-ServersTestJSON-657684837-project-member] Lock "f28d0043-236e-4df4-8478-e4b1befb6a7d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.680s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.398962] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bb6a5856-1d19-4f2e-989d-1545ecc1d8b1 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "6e289c20-8dd2-4680-b816-9bbe82fe7aec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.912s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.421652] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1091.421853] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1091.422053] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleting the datastore file [datastore2] 8cda6eea-d488-4d73-a7ab-7a905037929d {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1091.422595] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3e18aa75-6a78-42b2-b13a-33333d9976ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.428323] env[62952]: DEBUG oslo_vmware.api [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1091.428323] env[62952]: value = "task-1263873" [ 1091.428323] env[62952]: _type = "Task" [ 1091.428323] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.436247] env[62952]: DEBUG oslo_vmware.api [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263873, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.878806] env[62952]: DEBUG oslo_vmware.api [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263871, 'name': PowerOnVM_Task, 'duration_secs': 0.834878} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.879298] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1091.879598] env[62952]: INFO nova.compute.manager [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Took 7.42 seconds to spawn the instance on the hypervisor. [ 1091.879871] env[62952]: DEBUG nova.compute.manager [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1091.880771] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b20456-a27b-420c-9330-abb76cb75cb9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.942072] env[62952]: DEBUG oslo_vmware.api [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263873, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.442994} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.944139] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1091.944384] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1091.944613] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1091.944835] env[62952]: INFO nova.compute.manager [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1091.945170] env[62952]: DEBUG oslo.service.loopingcall [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1091.945656] env[62952]: DEBUG nova.compute.manager [-] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1091.945796] env[62952]: DEBUG nova.network.neutron [-] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1092.023204] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "42793459-89e6-4941-9b20-cbe52a241c6d" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.023467] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "42793459-89e6-4941-9b20-cbe52a241c6d" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.023646] env[62952]: INFO nova.compute.manager [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Shelving [ 1092.402790] env[62952]: INFO nova.compute.manager [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Took 12.26 seconds to build instance. [ 1092.537262] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1092.537593] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bebe0213-71ad-46a0-af1e-d6e7228cb199 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.545182] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1092.545182] env[62952]: value = "task-1263874" [ 1092.545182] env[62952]: _type = "Task" [ 1092.545182] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.556085] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263874, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.585263] env[62952]: DEBUG nova.compute.manager [req-bac531c2-2b17-4ca3-88e8-27cc5a9a2a3e req-b545a37d-2e46-4ba9-b3c2-c5f4fe126ee9 service nova] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Received event network-vif-deleted-116c22bf-b489-4039-9b75-598fe7be83dc {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1092.585603] env[62952]: INFO nova.compute.manager [req-bac531c2-2b17-4ca3-88e8-27cc5a9a2a3e req-b545a37d-2e46-4ba9-b3c2-c5f4fe126ee9 service nova] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Neutron deleted interface 116c22bf-b489-4039-9b75-598fe7be83dc; detaching it from the instance and deleting it from the info cache [ 1092.585841] env[62952]: DEBUG nova.network.neutron [req-bac531c2-2b17-4ca3-88e8-27cc5a9a2a3e req-b545a37d-2e46-4ba9-b3c2-c5f4fe126ee9 service nova] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.906542] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4634fe4c-0513-4abb-b549-15969024d425 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "126f57b5-be03-4164-b92f-3c2e96dd9171" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.773s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.988483] env[62952]: DEBUG oslo_concurrency.lockutils [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "43d79d26-4350-4329-bffc-cedfada40ee0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.989281] env[62952]: DEBUG oslo_concurrency.lockutils [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "43d79d26-4350-4329-bffc-cedfada40ee0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.059110] env[62952]: DEBUG nova.network.neutron [-] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.060395] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263874, 'name': PowerOffVM_Task, 'duration_secs': 0.19564} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.060654] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1093.061485] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c5ebc6-131f-4615-958e-6f37bb86388a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.084484] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5a2421-a883-4327-bafa-54d9c1366af3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.089705] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-04a8540f-2ac1-4621-a07f-9620551539c1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.114465] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-939dc15a-1902-4e49-b495-fedf292898b6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.143800] env[62952]: DEBUG nova.compute.manager [req-bac531c2-2b17-4ca3-88e8-27cc5a9a2a3e req-b545a37d-2e46-4ba9-b3c2-c5f4fe126ee9 service nova] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Detach interface failed, port_id=116c22bf-b489-4039-9b75-598fe7be83dc, reason: Instance 8cda6eea-d488-4d73-a7ab-7a905037929d could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1093.491436] env[62952]: DEBUG nova.compute.manager [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1093.566568] env[62952]: INFO nova.compute.manager [-] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Took 1.62 seconds to deallocate network for instance. [ 1093.601081] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1093.601523] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e5b7c3ac-ba78-4afa-af31-993b90b246f7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.608902] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1093.608902] env[62952]: value = "task-1263875" [ 1093.608902] env[62952]: _type = "Task" [ 1093.608902] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.617310] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263875, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.860026] env[62952]: INFO nova.compute.manager [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Rebuilding instance [ 1093.888108] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "126f57b5-be03-4164-b92f-3c2e96dd9171" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.888675] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "126f57b5-be03-4164-b92f-3c2e96dd9171" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.889235] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "126f57b5-be03-4164-b92f-3c2e96dd9171-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.889355] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "126f57b5-be03-4164-b92f-3c2e96dd9171-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.890030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "126f57b5-be03-4164-b92f-3c2e96dd9171-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.892797] env[62952]: INFO nova.compute.manager [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Terminating instance [ 1093.898049] env[62952]: DEBUG nova.compute.manager [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1093.898049] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1093.898536] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6625e59b-98a5-418a-bd1f-a7f7a74f652a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.915478] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1093.915849] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-439c9235-931b-489c-861d-03db144be149 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.923305] env[62952]: DEBUG oslo_vmware.api [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1093.923305] env[62952]: value = "task-1263876" [ 1093.923305] env[62952]: _type = "Task" [ 1093.923305] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.934935] env[62952]: DEBUG oslo_vmware.api [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263876, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.942567] env[62952]: DEBUG nova.compute.manager [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1093.943638] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-801b8f1b-e2ac-453d-a8d3-aaa1f73a72f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.030267] env[62952]: DEBUG oslo_concurrency.lockutils [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.030483] env[62952]: DEBUG oslo_concurrency.lockutils [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.032355] env[62952]: INFO nova.compute.claims [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1094.076413] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.120517] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263875, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.433226] env[62952]: DEBUG oslo_vmware.api [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263876, 'name': PowerOffVM_Task, 'duration_secs': 0.210062} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.433522] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1094.433697] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1094.433953] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6876b11f-360b-4efc-83ab-638754415be8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.457336] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1094.457729] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72ea2e92-e6b9-4bf7-b3f6-4ed49a37d254 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.463971] env[62952]: DEBUG oslo_vmware.api [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Waiting for the task: (returnval){ [ 1094.463971] env[62952]: value = "task-1263878" [ 1094.463971] env[62952]: _type = "Task" [ 1094.463971] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.472063] env[62952]: DEBUG oslo_vmware.api [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263878, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.496770] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1094.497129] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1094.497259] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleting the datastore file [datastore1] 126f57b5-be03-4164-b92f-3c2e96dd9171 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1094.497536] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-01de7d79-cbe0-465c-9386-1b38ea86b5c3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.504844] env[62952]: DEBUG oslo_vmware.api [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for the task: (returnval){ [ 1094.504844] env[62952]: value = "task-1263879" [ 1094.504844] env[62952]: _type = "Task" [ 1094.504844] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.513305] env[62952]: DEBUG oslo_vmware.api [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263879, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.619538] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263875, 'name': CreateSnapshot_Task, 'duration_secs': 0.73695} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.619929] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1094.620775] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94cbd0ec-c7b9-4591-833d-6a5669a91109 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.901806] env[62952]: DEBUG nova.compute.manager [req-48dcc617-8aab-4a3d-bba1-fa8cc614f763 req-95dc791d-195e-4794-897b-e02745674b52 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Received event network-changed-e1c235af-9501-47c0-a83a-ba40ea7fa8c1 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1094.902091] env[62952]: DEBUG nova.compute.manager [req-48dcc617-8aab-4a3d-bba1-fa8cc614f763 req-95dc791d-195e-4794-897b-e02745674b52 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Refreshing instance network info cache due to event network-changed-e1c235af-9501-47c0-a83a-ba40ea7fa8c1. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1094.902394] env[62952]: DEBUG oslo_concurrency.lockutils [req-48dcc617-8aab-4a3d-bba1-fa8cc614f763 req-95dc791d-195e-4794-897b-e02745674b52 service nova] Acquiring lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.902630] env[62952]: DEBUG oslo_concurrency.lockutils [req-48dcc617-8aab-4a3d-bba1-fa8cc614f763 req-95dc791d-195e-4794-897b-e02745674b52 service nova] Acquired lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.902887] env[62952]: DEBUG nova.network.neutron [req-48dcc617-8aab-4a3d-bba1-fa8cc614f763 req-95dc791d-195e-4794-897b-e02745674b52 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Refreshing network info cache for port e1c235af-9501-47c0-a83a-ba40ea7fa8c1 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1094.973665] env[62952]: DEBUG oslo_vmware.api [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263878, 'name': PowerOffVM_Task, 'duration_secs': 0.188471} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.974020] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1094.974731] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1094.975020] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-42781f87-288b-46fc-9821-9ad04fca15af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.982451] env[62952]: DEBUG oslo_vmware.api [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Waiting for the task: (returnval){ [ 1094.982451] env[62952]: value = "task-1263880" [ 1094.982451] env[62952]: _type = "Task" [ 1094.982451] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.991553] env[62952]: DEBUG oslo_vmware.api [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263880, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.014420] env[62952]: DEBUG oslo_vmware.api [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Task: {'id': task-1263879, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.289585} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.014682] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1095.014925] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1095.015060] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1095.015245] env[62952]: INFO nova.compute.manager [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1095.015487] env[62952]: DEBUG oslo.service.loopingcall [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1095.015686] env[62952]: DEBUG nova.compute.manager [-] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1095.015781] env[62952]: DEBUG nova.network.neutron [-] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1095.140057] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1095.142776] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0aa77fcb-fd50-44a0-a6da-9918c161aa50 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.151287] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1095.151287] env[62952]: value = "task-1263881" [ 1095.151287] env[62952]: _type = "Task" [ 1095.151287] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.162304] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263881, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.168342] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f58259-0e58-4758-b60c-5b9d549db179 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.175525] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49cd4e3e-2541-41b3-a2ec-8d80f4a90f3a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.204723] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5259f200-3c7a-4ab2-b4ae-42c2786e39d4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.211883] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-379add7d-7c73-4587-8e47-a3cbec2a8b80 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.224956] env[62952]: DEBUG nova.compute.provider_tree [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.493740] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1095.494059] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1095.494210] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271949', 'volume_id': '22582603-2be6-47fd-860e-60d0d1fd0f86', 'name': 'volume-22582603-2be6-47fd-860e-60d0d1fd0f86', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b', 'attached_at': '', 'detached_at': '', 'volume_id': '22582603-2be6-47fd-860e-60d0d1fd0f86', 'serial': '22582603-2be6-47fd-860e-60d0d1fd0f86'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1095.495017] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d704d0f1-bf86-4e6e-94dd-ced0596e7d28 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.516818] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-814cfc4c-345c-49b4-a7a9-4afa48fd786b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.523847] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f3d186-6678-4b4e-afe6-a054dc9203e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.546393] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e32657f-9b77-4165-bfca-460fcbf2a3cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.562345] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] The volume has not been displaced from its original location: [datastore2] volume-22582603-2be6-47fd-860e-60d0d1fd0f86/volume-22582603-2be6-47fd-860e-60d0d1fd0f86.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1095.567687] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Reconfiguring VM instance instance-0000005f to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1095.570105] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4dacacc9-c722-456c-afe6-c60c33c33806 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.588691] env[62952]: DEBUG oslo_vmware.api [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Waiting for the task: (returnval){ [ 1095.588691] env[62952]: value = "task-1263882" [ 1095.588691] env[62952]: _type = "Task" [ 1095.588691] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.596710] env[62952]: DEBUG oslo_vmware.api [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263882, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.661159] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263881, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.701801] env[62952]: DEBUG nova.network.neutron [req-48dcc617-8aab-4a3d-bba1-fa8cc614f763 req-95dc791d-195e-4794-897b-e02745674b52 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Updated VIF entry in instance network info cache for port e1c235af-9501-47c0-a83a-ba40ea7fa8c1. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1095.702166] env[62952]: DEBUG nova.network.neutron [req-48dcc617-8aab-4a3d-bba1-fa8cc614f763 req-95dc791d-195e-4794-897b-e02745674b52 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Updating instance_info_cache with network_info: [{"id": "e1c235af-9501-47c0-a83a-ba40ea7fa8c1", "address": "fa:16:3e:34:07:9b", "network": {"id": "827a8171-6759-4434-a6f1-437aad881c81", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1902486970-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091cdeb6da48492bb02f93822a45c9bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11b669be-fb26-4ef8-bdb6-c77ab9d06daf", "external-id": "nsx-vlan-transportzone-633", "segmentation_id": 633, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1c235af-95", "ovs_interfaceid": "e1c235af-9501-47c0-a83a-ba40ea7fa8c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.727972] env[62952]: DEBUG nova.scheduler.client.report [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1095.965897] env[62952]: DEBUG nova.network.neutron [-] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.099442] env[62952]: DEBUG oslo_vmware.api [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263882, 'name': ReconfigVM_Task, 'duration_secs': 0.179961} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.099653] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Reconfigured VM instance instance-0000005f to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1096.104798] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-352c4ca4-cd03-4672-b287-c2d8daf1707f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.121303] env[62952]: DEBUG oslo_vmware.api [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Waiting for the task: (returnval){ [ 1096.121303] env[62952]: value = "task-1263883" [ 1096.121303] env[62952]: _type = "Task" [ 1096.121303] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.129512] env[62952]: DEBUG oslo_vmware.api [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263883, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.161578] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263881, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.205463] env[62952]: DEBUG oslo_concurrency.lockutils [req-48dcc617-8aab-4a3d-bba1-fa8cc614f763 req-95dc791d-195e-4794-897b-e02745674b52 service nova] Releasing lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.205749] env[62952]: DEBUG nova.compute.manager [req-48dcc617-8aab-4a3d-bba1-fa8cc614f763 req-95dc791d-195e-4794-897b-e02745674b52 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Received event network-changed-e1c235af-9501-47c0-a83a-ba40ea7fa8c1 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1096.205926] env[62952]: DEBUG nova.compute.manager [req-48dcc617-8aab-4a3d-bba1-fa8cc614f763 req-95dc791d-195e-4794-897b-e02745674b52 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Refreshing instance network info cache due to event network-changed-e1c235af-9501-47c0-a83a-ba40ea7fa8c1. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1096.206166] env[62952]: DEBUG oslo_concurrency.lockutils [req-48dcc617-8aab-4a3d-bba1-fa8cc614f763 req-95dc791d-195e-4794-897b-e02745674b52 service nova] Acquiring lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.206319] env[62952]: DEBUG oslo_concurrency.lockutils [req-48dcc617-8aab-4a3d-bba1-fa8cc614f763 req-95dc791d-195e-4794-897b-e02745674b52 service nova] Acquired lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.206490] env[62952]: DEBUG nova.network.neutron [req-48dcc617-8aab-4a3d-bba1-fa8cc614f763 req-95dc791d-195e-4794-897b-e02745674b52 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Refreshing network info cache for port e1c235af-9501-47c0-a83a-ba40ea7fa8c1 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1096.233258] env[62952]: DEBUG oslo_concurrency.lockutils [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.203s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.233865] env[62952]: DEBUG nova.compute.manager [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1096.236707] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.161s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.236947] env[62952]: DEBUG nova.objects.instance [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lazy-loading 'resources' on Instance uuid 8cda6eea-d488-4d73-a7ab-7a905037929d {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1096.469074] env[62952]: INFO nova.compute.manager [-] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Took 1.45 seconds to deallocate network for instance. [ 1096.631817] env[62952]: DEBUG oslo_vmware.api [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263883, 'name': ReconfigVM_Task, 'duration_secs': 0.10111} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.632170] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271949', 'volume_id': '22582603-2be6-47fd-860e-60d0d1fd0f86', 'name': 'volume-22582603-2be6-47fd-860e-60d0d1fd0f86', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b', 'attached_at': '', 'detached_at': '', 'volume_id': '22582603-2be6-47fd-860e-60d0d1fd0f86', 'serial': '22582603-2be6-47fd-860e-60d0d1fd0f86'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1096.632433] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1096.633259] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48644f54-55e4-48bb-a158-0f84c5c96267 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.639480] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1096.639708] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-49c8719c-4541-46b4-b583-e713fb8cc14f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.660816] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263881, 'name': CloneVM_Task, 'duration_secs': 1.400506} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.661049] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Created linked-clone VM from snapshot [ 1096.661719] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72489768-dd6c-4013-a45a-dd8904c7d919 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.668816] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Uploading image 79448c63-26a7-44be-ba83-4d295285cf80 {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1096.693215] env[62952]: DEBUG oslo_vmware.rw_handles [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1096.693215] env[62952]: value = "vm-271973" [ 1096.693215] env[62952]: _type = "VirtualMachine" [ 1096.693215] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1096.693484] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4f17c9df-d772-47ac-bd19-017bbbb7a968 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.700872] env[62952]: DEBUG oslo_vmware.rw_handles [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lease: (returnval){ [ 1096.700872] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521f8373-9298-5b1b-a9c6-e00687995051" [ 1096.700872] env[62952]: _type = "HttpNfcLease" [ 1096.700872] env[62952]: } obtained for exporting VM: (result){ [ 1096.700872] env[62952]: value = "vm-271973" [ 1096.700872] env[62952]: _type = "VirtualMachine" [ 1096.700872] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1096.701143] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the lease: (returnval){ [ 1096.701143] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521f8373-9298-5b1b-a9c6-e00687995051" [ 1096.701143] env[62952]: _type = "HttpNfcLease" [ 1096.701143] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1096.704829] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1096.705051] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1096.705246] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Deleting the datastore file [datastore2] b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1096.705861] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-71b9701d-23a2-4511-a4d5-fed58e3086a1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.710801] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1096.710801] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521f8373-9298-5b1b-a9c6-e00687995051" [ 1096.710801] env[62952]: _type = "HttpNfcLease" [ 1096.710801] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1096.714742] env[62952]: DEBUG oslo_vmware.api [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Waiting for the task: (returnval){ [ 1096.714742] env[62952]: value = "task-1263886" [ 1096.714742] env[62952]: _type = "Task" [ 1096.714742] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.721553] env[62952]: DEBUG oslo_vmware.api [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263886, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.741776] env[62952]: DEBUG nova.compute.utils [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1096.746042] env[62952]: DEBUG nova.compute.manager [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1096.746042] env[62952]: DEBUG nova.network.neutron [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1096.791976] env[62952]: DEBUG nova.policy [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ffc122d64b19432aae8d9627d2928c06', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5080fbeac0744478843ab28efea8fc18', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1096.848269] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deccbb0d-20fe-4988-aed8-2fd0288d3e11 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.858096] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b17ffb5-e29a-45e9-972b-b14334db3a33 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.895218] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9981da5-fafb-4f89-ae8f-456916704d7e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.903483] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc1d857-1631-44e8-8bee-741c312892d1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.918372] env[62952]: DEBUG nova.compute.provider_tree [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1096.934275] env[62952]: DEBUG nova.compute.manager [req-9aedad98-8d73-4f49-8bcf-1603bdd22eb7 req-23473193-4aa2-43a7-bf0f-bd09ae4ac812 service nova] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Received event network-vif-deleted-7f9fb655-8880-45a0-9c96-75a3faf48d12 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1096.972260] env[62952]: DEBUG nova.network.neutron [req-48dcc617-8aab-4a3d-bba1-fa8cc614f763 req-95dc791d-195e-4794-897b-e02745674b52 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Updated VIF entry in instance network info cache for port e1c235af-9501-47c0-a83a-ba40ea7fa8c1. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1096.972689] env[62952]: DEBUG nova.network.neutron [req-48dcc617-8aab-4a3d-bba1-fa8cc614f763 req-95dc791d-195e-4794-897b-e02745674b52 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Updating instance_info_cache with network_info: [{"id": "e1c235af-9501-47c0-a83a-ba40ea7fa8c1", "address": "fa:16:3e:34:07:9b", "network": {"id": "827a8171-6759-4434-a6f1-437aad881c81", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1902486970-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091cdeb6da48492bb02f93822a45c9bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11b669be-fb26-4ef8-bdb6-c77ab9d06daf", "external-id": "nsx-vlan-transportzone-633", "segmentation_id": 633, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1c235af-95", "ovs_interfaceid": "e1c235af-9501-47c0-a83a-ba40ea7fa8c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.976687] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.103105] env[62952]: DEBUG nova.network.neutron [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Successfully created port: e6fdb64f-1d51-4b10-b297-9f760ae81c95 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1097.210847] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1097.210847] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521f8373-9298-5b1b-a9c6-e00687995051" [ 1097.210847] env[62952]: _type = "HttpNfcLease" [ 1097.210847] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1097.210847] env[62952]: DEBUG oslo_vmware.rw_handles [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1097.210847] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521f8373-9298-5b1b-a9c6-e00687995051" [ 1097.210847] env[62952]: _type = "HttpNfcLease" [ 1097.210847] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1097.210847] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f06d05-15c2-43e0-8318-13861d8d8c17 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.221886] env[62952]: DEBUG oslo_vmware.rw_handles [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ec13c7-af84-e9f5-e775-2d778ab56b86/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1097.222085] env[62952]: DEBUG oslo_vmware.rw_handles [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ec13c7-af84-e9f5-e775-2d778ab56b86/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1097.228592] env[62952]: DEBUG oslo_vmware.api [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Task: {'id': task-1263886, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196342} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.282705] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1097.282936] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1097.283130] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1097.285444] env[62952]: DEBUG nova.compute.manager [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1097.320550] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8efe4236-43cc-4698-833d-b7a58f1256be {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.350693] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1097.351305] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4b3f4c41-ba5a-41f0-81f5-e1261fd931e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.364805] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8403204-0336-42d2-b708-e9c738ed10de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.392040] env[62952]: ERROR nova.compute.manager [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Failed to detach volume 22582603-2be6-47fd-860e-60d0d1fd0f86 from /dev/sda: nova.exception.InstanceNotFound: Instance b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b could not be found. [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Traceback (most recent call last): [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] self.driver.rebuild(**kwargs) [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/virt/driver.py", line 390, in rebuild [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] raise NotImplementedError() [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] NotImplementedError [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] During handling of the above exception, another exception occurred: [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Traceback (most recent call last): [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] self.driver.detach_volume(context, old_connection_info, [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 558, in detach_volume [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] return self._volumeops.detach_volume(connection_info, instance) [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] self._detach_volume_vmdk(connection_info, instance) [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] stable_ref.fetch_moref(session) [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] nova.exception.InstanceNotFound: Instance b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b could not be found. [ 1097.392040] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] [ 1097.424174] env[62952]: DEBUG nova.scheduler.client.report [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1097.477393] env[62952]: DEBUG oslo_concurrency.lockutils [req-48dcc617-8aab-4a3d-bba1-fa8cc614f763 req-95dc791d-195e-4794-897b-e02745674b52 service nova] Releasing lock "refresh_cache-1941365e-d562-4d99-89eb-226a07e52071" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.536707] env[62952]: DEBUG nova.compute.utils [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Build of instance b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b aborted: Failed to rebuild volume backed instance. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1097.541188] env[62952]: ERROR nova.compute.manager [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b aborted: Failed to rebuild volume backed instance. [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Traceback (most recent call last): [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] self.driver.rebuild(**kwargs) [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/virt/driver.py", line 390, in rebuild [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] raise NotImplementedError() [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] NotImplementedError [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] During handling of the above exception, another exception occurred: [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Traceback (most recent call last): [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] self._detach_root_volume(context, instance, root_bdm) [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] with excutils.save_and_reraise_exception(): [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] self.force_reraise() [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] raise self.value [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] self.driver.detach_volume(context, old_connection_info, [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 558, in detach_volume [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] return self._volumeops.detach_volume(connection_info, instance) [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] self._detach_volume_vmdk(connection_info, instance) [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] stable_ref.fetch_moref(session) [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] nova.exception.InstanceNotFound: Instance b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b could not be found. [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] During handling of the above exception, another exception occurred: [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Traceback (most recent call last): [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/compute/manager.py", line 10863, in _error_out_instance_on_exception [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] yield [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 1097.541188] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] self._do_rebuild_instance_with_claim( [ 1097.542330] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 1097.542330] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] self._do_rebuild_instance( [ 1097.542330] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 1097.542330] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] self._rebuild_default_impl(**kwargs) [ 1097.542330] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 1097.542330] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] self._rebuild_volume_backed_instance( [ 1097.542330] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 1097.542330] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] raise exception.BuildAbortException( [ 1097.542330] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] nova.exception.BuildAbortException: Build of instance b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b aborted: Failed to rebuild volume backed instance. [ 1097.542330] env[62952]: ERROR nova.compute.manager [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] [ 1097.929056] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.692s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.932111] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.955s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.932438] env[62952]: DEBUG nova.objects.instance [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lazy-loading 'resources' on Instance uuid 126f57b5-be03-4164-b92f-3c2e96dd9171 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1097.961412] env[62952]: INFO nova.scheduler.client.report [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleted allocations for instance 8cda6eea-d488-4d73-a7ab-7a905037929d [ 1098.294687] env[62952]: DEBUG nova.compute.manager [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1098.319310] env[62952]: DEBUG nova.virt.hardware [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1098.319730] env[62952]: DEBUG nova.virt.hardware [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1098.319981] env[62952]: DEBUG nova.virt.hardware [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1098.320241] env[62952]: DEBUG nova.virt.hardware [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1098.320523] env[62952]: DEBUG nova.virt.hardware [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1098.320859] env[62952]: DEBUG nova.virt.hardware [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1098.321216] env[62952]: DEBUG nova.virt.hardware [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1098.321396] env[62952]: DEBUG nova.virt.hardware [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1098.321574] env[62952]: DEBUG nova.virt.hardware [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1098.321742] env[62952]: DEBUG nova.virt.hardware [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1098.321918] env[62952]: DEBUG nova.virt.hardware [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1098.322830] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d9ed92-66bd-459f-84aa-648745a915a3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.331417] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a26f6829-42ea-4b31-ba77-38bde98a8da5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.470805] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e7b23bc2-3116-4ce0-8ccc-42bbe3231bd3 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "8cda6eea-d488-4d73-a7ab-7a905037929d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.149s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.530400] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a21f68e-7ea7-465c-be79-5184c5d0a74f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.538882] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee38370f-305a-42e2-bff1-6cf14c2bd99d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.576037] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bb92e6c-2e2b-4d8b-98a6-c7b0b58d5f3e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.582038] env[62952]: DEBUG nova.compute.manager [req-34c7249d-0d4a-4ed7-a163-f81f20d4e208 req-711e9899-a623-4ca7-9c28-db1f043b4727 service nova] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Received event network-vif-plugged-e6fdb64f-1d51-4b10-b297-9f760ae81c95 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1098.582259] env[62952]: DEBUG oslo_concurrency.lockutils [req-34c7249d-0d4a-4ed7-a163-f81f20d4e208 req-711e9899-a623-4ca7-9c28-db1f043b4727 service nova] Acquiring lock "43d79d26-4350-4329-bffc-cedfada40ee0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.582899] env[62952]: DEBUG oslo_concurrency.lockutils [req-34c7249d-0d4a-4ed7-a163-f81f20d4e208 req-711e9899-a623-4ca7-9c28-db1f043b4727 service nova] Lock "43d79d26-4350-4329-bffc-cedfada40ee0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.583300] env[62952]: DEBUG oslo_concurrency.lockutils [req-34c7249d-0d4a-4ed7-a163-f81f20d4e208 req-711e9899-a623-4ca7-9c28-db1f043b4727 service nova] Lock "43d79d26-4350-4329-bffc-cedfada40ee0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.583489] env[62952]: DEBUG nova.compute.manager [req-34c7249d-0d4a-4ed7-a163-f81f20d4e208 req-711e9899-a623-4ca7-9c28-db1f043b4727 service nova] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] No waiting events found dispatching network-vif-plugged-e6fdb64f-1d51-4b10-b297-9f760ae81c95 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1098.583670] env[62952]: WARNING nova.compute.manager [req-34c7249d-0d4a-4ed7-a163-f81f20d4e208 req-711e9899-a623-4ca7-9c28-db1f043b4727 service nova] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Received unexpected event network-vif-plugged-e6fdb64f-1d51-4b10-b297-9f760ae81c95 for instance with vm_state building and task_state spawning. [ 1098.590840] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d8989c-182f-422b-a1c1-b99c3b32e41e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.607420] env[62952]: DEBUG nova.compute.provider_tree [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1099.109771] env[62952]: DEBUG nova.scheduler.client.report [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1099.174049] env[62952]: DEBUG nova.network.neutron [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Successfully updated port: e6fdb64f-1d51-4b10-b297-9f760ae81c95 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1099.195811] env[62952]: DEBUG nova.compute.manager [req-f96d2ed8-a13d-4969-b39d-452d4e32e3b4 req-978d26b7-90d2-4a37-bb20-307d3d173cd4 service nova] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Received event network-changed-e6fdb64f-1d51-4b10-b297-9f760ae81c95 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1099.196068] env[62952]: DEBUG nova.compute.manager [req-f96d2ed8-a13d-4969-b39d-452d4e32e3b4 req-978d26b7-90d2-4a37-bb20-307d3d173cd4 service nova] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Refreshing instance network info cache due to event network-changed-e6fdb64f-1d51-4b10-b297-9f760ae81c95. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1099.196216] env[62952]: DEBUG oslo_concurrency.lockutils [req-f96d2ed8-a13d-4969-b39d-452d4e32e3b4 req-978d26b7-90d2-4a37-bb20-307d3d173cd4 service nova] Acquiring lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1099.196400] env[62952]: DEBUG oslo_concurrency.lockutils [req-f96d2ed8-a13d-4969-b39d-452d4e32e3b4 req-978d26b7-90d2-4a37-bb20-307d3d173cd4 service nova] Acquired lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.196583] env[62952]: DEBUG nova.network.neutron [req-f96d2ed8-a13d-4969-b39d-452d4e32e3b4 req-978d26b7-90d2-4a37-bb20-307d3d173cd4 service nova] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Refreshing network info cache for port e6fdb64f-1d51-4b10-b297-9f760ae81c95 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1099.581451] env[62952]: DEBUG oslo_concurrency.lockutils [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.589448] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.589716] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.615181] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.683s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.618106] env[62952]: DEBUG oslo_concurrency.lockutils [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.037s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.633586] env[62952]: INFO nova.scheduler.client.report [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Deleted allocations for instance 126f57b5-be03-4164-b92f-3c2e96dd9171 [ 1099.678043] env[62952]: DEBUG oslo_concurrency.lockutils [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1099.715517] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c90aa4f-b17c-4f7c-8aa9-f29f2ac65c80 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.724175] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf11f055-74d8-405d-82c0-8540308722fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.757950] env[62952]: DEBUG nova.network.neutron [req-f96d2ed8-a13d-4969-b39d-452d4e32e3b4 req-978d26b7-90d2-4a37-bb20-307d3d173cd4 service nova] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1099.760445] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac66aa7-0e53-4f05-a26d-1545d077c58b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.768518] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f167208f-94a9-450c-bc55-1b94e251ceb5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.783134] env[62952]: DEBUG nova.compute.provider_tree [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1099.845084] env[62952]: DEBUG nova.network.neutron [req-f96d2ed8-a13d-4969-b39d-452d4e32e3b4 req-978d26b7-90d2-4a37-bb20-307d3d173cd4 service nova] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.071714] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Acquiring lock "b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.072035] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Lock "b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.072289] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Acquiring lock "b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.072506] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Lock "b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.072687] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Lock "b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.074953] env[62952]: INFO nova.compute.manager [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Terminating instance [ 1100.077221] env[62952]: DEBUG nova.compute.manager [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1100.077607] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5dc22110-f0cf-4008-9460-8da69050a304 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.087054] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa194fc-6b73-45aa-9c53-afae8bfbd2ed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.097657] env[62952]: DEBUG nova.compute.manager [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1100.116199] env[62952]: WARNING nova.virt.vmwareapi.driver [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b could not be found. [ 1100.116511] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1100.116742] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7b3fa104-0e74-4e0b-b599-ecfda76b4282 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.125960] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9a6b8e-26f7-4fdf-9ea6-d77703f0a41a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.143262] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d0801a-5aaf-447d-a512-1b75d3614a27 tempest-ServerDiskConfigTestJSON-1471667218 tempest-ServerDiskConfigTestJSON-1471667218-project-member] Lock "126f57b5-be03-4164-b92f-3c2e96dd9171" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.255s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.154905] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b could not be found. [ 1100.155132] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1100.155355] env[62952]: INFO nova.compute.manager [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Took 0.08 seconds to destroy the instance on the hypervisor. [ 1100.155611] env[62952]: DEBUG oslo.service.loopingcall [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1100.155833] env[62952]: DEBUG nova.compute.manager [-] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1100.155931] env[62952]: DEBUG nova.network.neutron [-] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1100.285970] env[62952]: DEBUG nova.scheduler.client.report [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1100.347743] env[62952]: DEBUG oslo_concurrency.lockutils [req-f96d2ed8-a13d-4969-b39d-452d4e32e3b4 req-978d26b7-90d2-4a37-bb20-307d3d173cd4 service nova] Releasing lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.348164] env[62952]: DEBUG oslo_concurrency.lockutils [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.348362] env[62952]: DEBUG nova.network.neutron [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1100.623470] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.791882] env[62952]: DEBUG oslo_concurrency.lockutils [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.174s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.792320] env[62952]: INFO nova.compute.manager [None req-abccaede-a7d8-41a7-b75f-025b8f754f3d tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Successfully reverted task state from rebuilding on failure for instance. [ 1100.797954] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.175s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.799720] env[62952]: INFO nova.compute.claims [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1100.928831] env[62952]: DEBUG nova.network.neutron [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1101.131482] env[62952]: DEBUG nova.network.neutron [-] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.203624] env[62952]: DEBUG nova.network.neutron [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating instance_info_cache with network_info: [{"id": "e6fdb64f-1d51-4b10-b297-9f760ae81c95", "address": "fa:16:3e:95:ff:4c", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6fdb64f-1d", "ovs_interfaceid": "e6fdb64f-1d51-4b10-b297-9f760ae81c95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.228613] env[62952]: DEBUG nova.compute.manager [req-ea39cbdb-5622-4710-a680-9a3e0e6ccfc5 req-86685698-d089-4714-b865-e85e172d2914 service nova] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Received event network-vif-deleted-d5013042-c6e4-4118-aaa6-3947b89a6fb3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1101.634578] env[62952]: INFO nova.compute.manager [-] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Took 1.48 seconds to deallocate network for instance. [ 1101.706300] env[62952]: DEBUG oslo_concurrency.lockutils [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.706679] env[62952]: DEBUG nova.compute.manager [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Instance network_info: |[{"id": "e6fdb64f-1d51-4b10-b297-9f760ae81c95", "address": "fa:16:3e:95:ff:4c", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6fdb64f-1d", "ovs_interfaceid": "e6fdb64f-1d51-4b10-b297-9f760ae81c95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1101.707158] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:ff:4c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e1c9bb98-73a9-48eb-856e-a541afe9b07b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e6fdb64f-1d51-4b10-b297-9f760ae81c95', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1101.716927] env[62952]: DEBUG oslo.service.loopingcall [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1101.717255] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1101.717485] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2c74a7fc-ed53-46fb-8539-ac72a45d8efb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.740822] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1101.740822] env[62952]: value = "task-1263887" [ 1101.740822] env[62952]: _type = "Task" [ 1101.740822] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.751538] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263887, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.907915] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acddf535-5a4c-4ca1-a466-b89f3511ad54 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.915808] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35186558-8437-485a-9c73-400d3967c97b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.954301] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ff15d5-eeb0-4071-9d1f-28fb2431ab9d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.964066] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35902203-72c9-416d-bd5b-501978ee5469 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.978209] env[62952]: DEBUG nova.compute.provider_tree [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1102.196702] env[62952]: INFO nova.compute.manager [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Took 0.56 seconds to detach 1 volumes for instance. [ 1102.201022] env[62952]: DEBUG nova.compute.manager [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Deleting volume: 22582603-2be6-47fd-860e-60d0d1fd0f86 {{(pid=62952) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1102.252201] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263887, 'name': CreateVM_Task, 'duration_secs': 0.466602} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.256352] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1102.256704] env[62952]: DEBUG oslo_concurrency.lockutils [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.256883] env[62952]: DEBUG oslo_concurrency.lockutils [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.257277] env[62952]: DEBUG oslo_concurrency.lockutils [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1102.257772] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c14b8c94-51e2-4f47-8538-8c1fddfc2a1c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.263407] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1102.263407] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d55b62-3b7b-d0c6-0bc1-1c0d77d3d5d0" [ 1102.263407] env[62952]: _type = "Task" [ 1102.263407] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.271869] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d55b62-3b7b-d0c6-0bc1-1c0d77d3d5d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.481715] env[62952]: DEBUG nova.scheduler.client.report [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1102.758990] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.774099] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d55b62-3b7b-d0c6-0bc1-1c0d77d3d5d0, 'name': SearchDatastore_Task, 'duration_secs': 0.013033} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.774447] env[62952]: DEBUG oslo_concurrency.lockutils [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1102.774761] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1102.775061] env[62952]: DEBUG oslo_concurrency.lockutils [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.775587] env[62952]: DEBUG oslo_concurrency.lockutils [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.775587] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1102.775728] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e45ce07d-1bec-4c72-9308-dcdccb708a18 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.785157] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1102.785400] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1102.786187] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2461a7ce-463b-4f12-a4de-bc2182974c08 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.792687] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1102.792687] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b9e348-b7b1-7fdf-3a9a-4d4b21f23545" [ 1102.792687] env[62952]: _type = "Task" [ 1102.792687] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.801365] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b9e348-b7b1-7fdf-3a9a-4d4b21f23545, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.987777] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.190s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.988399] env[62952]: DEBUG nova.compute.manager [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1102.992274] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.233s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.992274] env[62952]: DEBUG nova.objects.instance [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Lazy-loading 'resources' on Instance uuid b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1103.307127] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52b9e348-b7b1-7fdf-3a9a-4d4b21f23545, 'name': SearchDatastore_Task, 'duration_secs': 0.011867} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.308081] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4ed83d0-ed4d-41d5-a1d5-0c956d3c8ab7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.313924] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1103.313924] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f9b6aa-217b-dc65-8e37-fba5c045b0cf" [ 1103.313924] env[62952]: _type = "Task" [ 1103.313924] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.324069] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f9b6aa-217b-dc65-8e37-fba5c045b0cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.497311] env[62952]: DEBUG nova.compute.utils [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1103.501806] env[62952]: DEBUG nova.compute.manager [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1103.501989] env[62952]: DEBUG nova.network.neutron [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1103.579911] env[62952]: DEBUG nova.policy [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '469eae9db3d14e9f90988e3093514e69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff928791a3cf496cb8db2c2a6c47a6a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1103.654362] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de53e5e2-7aec-4d9a-9a39-bd4a34ecd084 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.663241] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee05730-9401-4318-8984-44cb7d37de94 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.693693] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d4c42da-911b-4d5e-9275-7e54560005e7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.701310] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baaea07d-6800-45b3-8638-fe044b9b75a1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.714766] env[62952]: DEBUG nova.compute.provider_tree [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1103.826751] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f9b6aa-217b-dc65-8e37-fba5c045b0cf, 'name': SearchDatastore_Task, 'duration_secs': 0.009429} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.827224] env[62952]: DEBUG oslo_concurrency.lockutils [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.827555] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 43d79d26-4350-4329-bffc-cedfada40ee0/43d79d26-4350-4329-bffc-cedfada40ee0.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1103.827869] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-493ced6c-8894-4a91-8cd3-2f698269e458 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.834649] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1103.834649] env[62952]: value = "task-1263889" [ 1103.834649] env[62952]: _type = "Task" [ 1103.834649] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.843584] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263889, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.005640] env[62952]: DEBUG nova.compute.manager [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1104.218281] env[62952]: DEBUG nova.scheduler.client.report [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1104.227323] env[62952]: DEBUG nova.network.neutron [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Successfully created port: 9d5ab9ea-3d2a-40d3-996d-96b27d38cff5 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1104.361715] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263889, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.723083] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.731s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.742956] env[62952]: DEBUG oslo_vmware.rw_handles [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ec13c7-af84-e9f5-e775-2d778ab56b86/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1104.744016] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082fc019-dcbd-4f5d-bd1f-f821ef753b4c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.751685] env[62952]: DEBUG oslo_vmware.rw_handles [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ec13c7-af84-e9f5-e775-2d778ab56b86/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1104.752531] env[62952]: ERROR oslo_vmware.rw_handles [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ec13c7-af84-e9f5-e775-2d778ab56b86/disk-0.vmdk due to incomplete transfer. [ 1104.752841] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-8e212931-e4de-41d6-8a60-ea4917c77a3d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.762716] env[62952]: DEBUG oslo_vmware.rw_handles [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ec13c7-af84-e9f5-e775-2d778ab56b86/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1104.762992] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Uploaded image 79448c63-26a7-44be-ba83-4d295285cf80 to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1104.766653] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1104.766981] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-36577ddd-dad5-480f-8c6d-295dedebb123 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.775930] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1104.775930] env[62952]: value = "task-1263890" [ 1104.775930] env[62952]: _type = "Task" [ 1104.775930] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.786376] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263890, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.847811] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263889, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.60723} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.848269] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 43d79d26-4350-4329-bffc-cedfada40ee0/43d79d26-4350-4329-bffc-cedfada40ee0.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1104.848535] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1104.848835] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5f000619-816d-4aaa-aae6-ad8b6cd9f1f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.857071] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1104.857071] env[62952]: value = "task-1263891" [ 1104.857071] env[62952]: _type = "Task" [ 1104.857071] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.872105] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263891, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.015613] env[62952]: DEBUG nova.compute.manager [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1105.042792] env[62952]: DEBUG nova.virt.hardware [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1105.043084] env[62952]: DEBUG nova.virt.hardware [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1105.043256] env[62952]: DEBUG nova.virt.hardware [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1105.043448] env[62952]: DEBUG nova.virt.hardware [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1105.043699] env[62952]: DEBUG nova.virt.hardware [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1105.043876] env[62952]: DEBUG nova.virt.hardware [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1105.044232] env[62952]: DEBUG nova.virt.hardware [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1105.044494] env[62952]: DEBUG nova.virt.hardware [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1105.044756] env[62952]: DEBUG nova.virt.hardware [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1105.045016] env[62952]: DEBUG nova.virt.hardware [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1105.045306] env[62952]: DEBUG nova.virt.hardware [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1105.046643] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-794ea122-cf71-48ae-a2c2-2c2cfa5f0a65 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.056198] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b986b8c0-acc6-45f3-a7a8-6b3805ea4259 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.252315] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2e945b82-28a4-4531-a5cd-1df4e396a740 tempest-ServerActionsV293TestJSON-947201152 tempest-ServerActionsV293TestJSON-947201152-project-member] Lock "b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.179s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.289666] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263890, 'name': Destroy_Task, 'duration_secs': 0.326994} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.290016] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Destroyed the VM [ 1105.290501] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1105.290798] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5653aae6-49f3-4d95-96c7-5e03e47c5899 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.298456] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1105.298456] env[62952]: value = "task-1263892" [ 1105.298456] env[62952]: _type = "Task" [ 1105.298456] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.308125] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263892, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.367513] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263891, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067283} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.367827] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1105.368664] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819a1854-da9a-4e1c-9eb3-2eb1c929f0a9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.401781] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 43d79d26-4350-4329-bffc-cedfada40ee0/43d79d26-4350-4329-bffc-cedfada40ee0.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1105.402318] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c40cb089-d007-4fb0-9085-5fb623a18b39 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.425477] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1105.425477] env[62952]: value = "task-1263893" [ 1105.425477] env[62952]: _type = "Task" [ 1105.425477] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.434837] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263893, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.809154] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263892, 'name': RemoveSnapshot_Task, 'duration_secs': 0.364423} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.809421] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1105.809695] env[62952]: DEBUG nova.compute.manager [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1105.810548] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61fd7e78-e1f7-4155-91df-6aa549928b45 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.940457] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263893, 'name': ReconfigVM_Task, 'duration_secs': 0.303833} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.942040] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 43d79d26-4350-4329-bffc-cedfada40ee0/43d79d26-4350-4329-bffc-cedfada40ee0.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1105.942040] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-174d6f90-7ccd-48ea-a7de-7bdfb627e0df {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.948218] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1105.948218] env[62952]: value = "task-1263894" [ 1105.948218] env[62952]: _type = "Task" [ 1105.948218] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.957777] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263894, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.324945] env[62952]: INFO nova.compute.manager [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Shelve offloading [ 1106.328992] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1106.328992] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e2fc313b-c78f-49f8-8da8-0bca02ab2d3e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.335032] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1106.335032] env[62952]: value = "task-1263895" [ 1106.335032] env[62952]: _type = "Task" [ 1106.335032] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.344016] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1106.344616] env[62952]: DEBUG nova.compute.manager [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1106.345433] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a3bf41-b3b4-41f9-8650-012da3a5b58d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.353124] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.353315] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.353495] env[62952]: DEBUG nova.network.neutron [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1106.436347] env[62952]: DEBUG nova.compute.manager [req-32137193-9564-4653-b941-b1ebdf9797df req-ddf0c4d7-db90-4127-afa2-155ed33a893a service nova] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Received event network-vif-plugged-9d5ab9ea-3d2a-40d3-996d-96b27d38cff5 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1106.436605] env[62952]: DEBUG oslo_concurrency.lockutils [req-32137193-9564-4653-b941-b1ebdf9797df req-ddf0c4d7-db90-4127-afa2-155ed33a893a service nova] Acquiring lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.436774] env[62952]: DEBUG oslo_concurrency.lockutils [req-32137193-9564-4653-b941-b1ebdf9797df req-ddf0c4d7-db90-4127-afa2-155ed33a893a service nova] Lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.436960] env[62952]: DEBUG oslo_concurrency.lockutils [req-32137193-9564-4653-b941-b1ebdf9797df req-ddf0c4d7-db90-4127-afa2-155ed33a893a service nova] Lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.437202] env[62952]: DEBUG nova.compute.manager [req-32137193-9564-4653-b941-b1ebdf9797df req-ddf0c4d7-db90-4127-afa2-155ed33a893a service nova] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] No waiting events found dispatching network-vif-plugged-9d5ab9ea-3d2a-40d3-996d-96b27d38cff5 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1106.437388] env[62952]: WARNING nova.compute.manager [req-32137193-9564-4653-b941-b1ebdf9797df req-ddf0c4d7-db90-4127-afa2-155ed33a893a service nova] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Received unexpected event network-vif-plugged-9d5ab9ea-3d2a-40d3-996d-96b27d38cff5 for instance with vm_state building and task_state spawning. [ 1106.458319] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263894, 'name': Rename_Task, 'duration_secs': 0.371145} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.458781] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1106.459658] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bda3cafd-e9b0-4160-afd4-e88cca0be44f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.466479] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1106.466479] env[62952]: value = "task-1263896" [ 1106.466479] env[62952]: _type = "Task" [ 1106.466479] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.473985] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263896, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.680725] env[62952]: DEBUG nova.network.neutron [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Successfully updated port: 9d5ab9ea-3d2a-40d3-996d-96b27d38cff5 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1106.977200] env[62952]: DEBUG oslo_vmware.api [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263896, 'name': PowerOnVM_Task, 'duration_secs': 0.435163} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.977494] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1106.977723] env[62952]: INFO nova.compute.manager [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Took 8.68 seconds to spawn the instance on the hypervisor. [ 1106.977888] env[62952]: DEBUG nova.compute.manager [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1106.978753] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d0396c-7521-41d1-8e0f-890caba29554 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.185642] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "refresh_cache-11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.185642] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "refresh_cache-11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.185642] env[62952]: DEBUG nova.network.neutron [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1107.258971] env[62952]: DEBUG nova.network.neutron [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Updating instance_info_cache with network_info: [{"id": "ddd93d49-eb57-4af9-a9bb-35d8269b939d", "address": "fa:16:3e:e3:8d:93", "network": {"id": "247a88e9-e8e8-46af-ac41-6777f0782fe0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1297422061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.226", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d93803f89ca64e66b8271e7b0bb48983", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddd93d49-eb", "ovs_interfaceid": "ddd93d49-eb57-4af9-a9bb-35d8269b939d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.495702] env[62952]: INFO nova.compute.manager [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Took 13.50 seconds to build instance. [ 1107.641369] env[62952]: DEBUG oslo_concurrency.lockutils [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "f92ef782-fd64-47f0-8145-9261a1c3005c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.642053] env[62952]: DEBUG oslo_concurrency.lockutils [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "f92ef782-fd64-47f0-8145-9261a1c3005c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.735449] env[62952]: DEBUG nova.network.neutron [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1107.764333] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Releasing lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.967974] env[62952]: DEBUG nova.network.neutron [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Updating instance_info_cache with network_info: [{"id": "9d5ab9ea-3d2a-40d3-996d-96b27d38cff5", "address": "fa:16:3e:4e:7f:ef", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d5ab9ea-3d", "ovs_interfaceid": "9d5ab9ea-3d2a-40d3-996d-96b27d38cff5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.999965] env[62952]: DEBUG oslo_concurrency.lockutils [None req-68a54af3-f4e1-4529-a8a5-7c3bb8a12c30 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "43d79d26-4350-4329-bffc-cedfada40ee0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.011s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.022449] env[62952]: DEBUG nova.compute.manager [req-55559970-7a60-4cc9-a2d1-2f43161c53ac req-786695ae-12b0-4eb7-a48a-c1fd74ed181d service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Received event network-vif-unplugged-ddd93d49-eb57-4af9-a9bb-35d8269b939d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1108.022815] env[62952]: DEBUG oslo_concurrency.lockutils [req-55559970-7a60-4cc9-a2d1-2f43161c53ac req-786695ae-12b0-4eb7-a48a-c1fd74ed181d service nova] Acquiring lock "42793459-89e6-4941-9b20-cbe52a241c6d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.023114] env[62952]: DEBUG oslo_concurrency.lockutils [req-55559970-7a60-4cc9-a2d1-2f43161c53ac req-786695ae-12b0-4eb7-a48a-c1fd74ed181d service nova] Lock "42793459-89e6-4941-9b20-cbe52a241c6d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.023370] env[62952]: DEBUG oslo_concurrency.lockutils [req-55559970-7a60-4cc9-a2d1-2f43161c53ac req-786695ae-12b0-4eb7-a48a-c1fd74ed181d service nova] Lock "42793459-89e6-4941-9b20-cbe52a241c6d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.023589] env[62952]: DEBUG nova.compute.manager [req-55559970-7a60-4cc9-a2d1-2f43161c53ac req-786695ae-12b0-4eb7-a48a-c1fd74ed181d service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] No waiting events found dispatching network-vif-unplugged-ddd93d49-eb57-4af9-a9bb-35d8269b939d {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1108.023807] env[62952]: WARNING nova.compute.manager [req-55559970-7a60-4cc9-a2d1-2f43161c53ac req-786695ae-12b0-4eb7-a48a-c1fd74ed181d service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Received unexpected event network-vif-unplugged-ddd93d49-eb57-4af9-a9bb-35d8269b939d for instance with vm_state shelved and task_state shelving_offloading. [ 1108.103937] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1108.104978] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5512afe-5e3b-4f14-9f5a-17568bbba339 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.117428] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1108.118201] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc415b5a-d07e-4e51-b2d0-2ad0bf78f6ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.144839] env[62952]: DEBUG nova.compute.manager [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1108.178928] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1108.181668] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1108.181999] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Deleting the datastore file [datastore2] 42793459-89e6-4941-9b20-cbe52a241c6d {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1108.183332] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-27bd7ae3-9fe0-4cba-bd16-c7cbdc09a215 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.191905] env[62952]: DEBUG oslo_concurrency.lockutils [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Acquiring lock "148e52e8-07ef-4b3d-a43b-eefc6820b248" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.192149] env[62952]: DEBUG oslo_concurrency.lockutils [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Lock "148e52e8-07ef-4b3d-a43b-eefc6820b248" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.197936] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1108.197936] env[62952]: value = "task-1263898" [ 1108.197936] env[62952]: _type = "Task" [ 1108.197936] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.206808] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263898, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.470948] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "refresh_cache-11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.471257] env[62952]: DEBUG nova.compute.manager [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Instance network_info: |[{"id": "9d5ab9ea-3d2a-40d3-996d-96b27d38cff5", "address": "fa:16:3e:4e:7f:ef", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d5ab9ea-3d", "ovs_interfaceid": "9d5ab9ea-3d2a-40d3-996d-96b27d38cff5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1108.471686] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4e:7f:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9d5ab9ea-3d2a-40d3-996d-96b27d38cff5', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1108.481099] env[62952]: DEBUG oslo.service.loopingcall [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1108.482503] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1108.483927] env[62952]: DEBUG nova.compute.manager [req-08de22cf-0829-4ffe-b5f5-12e9b4198ecc req-c4207f99-8397-4dfc-957b-9add08d4243d service nova] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Received event network-changed-9d5ab9ea-3d2a-40d3-996d-96b27d38cff5 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1108.484234] env[62952]: DEBUG nova.compute.manager [req-08de22cf-0829-4ffe-b5f5-12e9b4198ecc req-c4207f99-8397-4dfc-957b-9add08d4243d service nova] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Refreshing instance network info cache due to event network-changed-9d5ab9ea-3d2a-40d3-996d-96b27d38cff5. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1108.484604] env[62952]: DEBUG oslo_concurrency.lockutils [req-08de22cf-0829-4ffe-b5f5-12e9b4198ecc req-c4207f99-8397-4dfc-957b-9add08d4243d service nova] Acquiring lock "refresh_cache-11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.485483] env[62952]: DEBUG oslo_concurrency.lockutils [req-08de22cf-0829-4ffe-b5f5-12e9b4198ecc req-c4207f99-8397-4dfc-957b-9add08d4243d service nova] Acquired lock "refresh_cache-11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.485483] env[62952]: DEBUG nova.network.neutron [req-08de22cf-0829-4ffe-b5f5-12e9b4198ecc req-c4207f99-8397-4dfc-957b-9add08d4243d service nova] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Refreshing network info cache for port 9d5ab9ea-3d2a-40d3-996d-96b27d38cff5 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1108.486600] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-32fd8705-4b79-43c4-a577-36b4669af3f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.509440] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1108.509440] env[62952]: value = "task-1263899" [ 1108.509440] env[62952]: _type = "Task" [ 1108.509440] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.518616] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263899, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.672632] env[62952]: DEBUG oslo_concurrency.lockutils [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.672935] env[62952]: DEBUG oslo_concurrency.lockutils [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.674691] env[62952]: INFO nova.compute.claims [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1108.697622] env[62952]: DEBUG nova.compute.manager [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1108.717916] env[62952]: DEBUG oslo_vmware.api [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263898, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130312} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.718417] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1108.718417] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1108.718606] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1108.754224] env[62952]: INFO nova.scheduler.client.report [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Deleted allocations for instance 42793459-89e6-4941-9b20-cbe52a241c6d [ 1109.019256] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263899, 'name': CreateVM_Task, 'duration_secs': 0.298964} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.019434] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1109.020181] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.020378] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.020705] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1109.020973] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9109951-161f-4fbc-9165-d6f4227aa9b1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.025914] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1109.025914] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5287e421-a4a6-a237-ffd9-e70594cefef0" [ 1109.025914] env[62952]: _type = "Task" [ 1109.025914] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.033825] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5287e421-a4a6-a237-ffd9-e70594cefef0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.222688] env[62952]: DEBUG oslo_concurrency.lockutils [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.227590] env[62952]: DEBUG nova.network.neutron [req-08de22cf-0829-4ffe-b5f5-12e9b4198ecc req-c4207f99-8397-4dfc-957b-9add08d4243d service nova] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Updated VIF entry in instance network info cache for port 9d5ab9ea-3d2a-40d3-996d-96b27d38cff5. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1109.227979] env[62952]: DEBUG nova.network.neutron [req-08de22cf-0829-4ffe-b5f5-12e9b4198ecc req-c4207f99-8397-4dfc-957b-9add08d4243d service nova] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Updating instance_info_cache with network_info: [{"id": "9d5ab9ea-3d2a-40d3-996d-96b27d38cff5", "address": "fa:16:3e:4e:7f:ef", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d5ab9ea-3d", "ovs_interfaceid": "9d5ab9ea-3d2a-40d3-996d-96b27d38cff5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.259095] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.537181] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5287e421-a4a6-a237-ffd9-e70594cefef0, 'name': SearchDatastore_Task, 'duration_secs': 0.009857} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.537536] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.537789] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1109.538051] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.538212] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.538404] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1109.538694] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-56167526-a558-4fec-98f8-29ee7ca8b2da {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.547550] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1109.547900] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1109.548590] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77ac75a4-93e5-41eb-a9a0-2872470b3e4b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.555050] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1109.555050] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5202698a-a639-4f0b-07ce-d5fade8ca9ff" [ 1109.555050] env[62952]: _type = "Task" [ 1109.555050] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.564261] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5202698a-a639-4f0b-07ce-d5fade8ca9ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.731259] env[62952]: DEBUG oslo_concurrency.lockutils [req-08de22cf-0829-4ffe-b5f5-12e9b4198ecc req-c4207f99-8397-4dfc-957b-9add08d4243d service nova] Releasing lock "refresh_cache-11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.731732] env[62952]: DEBUG nova.compute.manager [req-08de22cf-0829-4ffe-b5f5-12e9b4198ecc req-c4207f99-8397-4dfc-957b-9add08d4243d service nova] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Received event network-changed-e6fdb64f-1d51-4b10-b297-9f760ae81c95 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1109.732020] env[62952]: DEBUG nova.compute.manager [req-08de22cf-0829-4ffe-b5f5-12e9b4198ecc req-c4207f99-8397-4dfc-957b-9add08d4243d service nova] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Refreshing instance network info cache due to event network-changed-e6fdb64f-1d51-4b10-b297-9f760ae81c95. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1109.732429] env[62952]: DEBUG oslo_concurrency.lockutils [req-08de22cf-0829-4ffe-b5f5-12e9b4198ecc req-c4207f99-8397-4dfc-957b-9add08d4243d service nova] Acquiring lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.732685] env[62952]: DEBUG oslo_concurrency.lockutils [req-08de22cf-0829-4ffe-b5f5-12e9b4198ecc req-c4207f99-8397-4dfc-957b-9add08d4243d service nova] Acquired lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.732884] env[62952]: DEBUG nova.network.neutron [req-08de22cf-0829-4ffe-b5f5-12e9b4198ecc req-c4207f99-8397-4dfc-957b-9add08d4243d service nova] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Refreshing network info cache for port e6fdb64f-1d51-4b10-b297-9f760ae81c95 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1109.798511] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0d0b6ae-5ab5-4405-b5c4-d88fa57b0498 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.807854] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3394a740-34a3-43de-9d19-0b53f9dc413b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.850304] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e1e8c13-8dd1-4b33-a870-26e4af72427d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.857988] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3375d07d-299c-44be-a4c6-e572f6a8b48a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.873062] env[62952]: DEBUG nova.compute.provider_tree [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1110.061884] env[62952]: DEBUG nova.compute.manager [req-d590c8c1-282b-4433-a599-89ee6842eb94 req-52776f0b-e393-41a6-b940-637af4b64cba service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Received event network-changed-ddd93d49-eb57-4af9-a9bb-35d8269b939d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1110.062056] env[62952]: DEBUG nova.compute.manager [req-d590c8c1-282b-4433-a599-89ee6842eb94 req-52776f0b-e393-41a6-b940-637af4b64cba service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Refreshing instance network info cache due to event network-changed-ddd93d49-eb57-4af9-a9bb-35d8269b939d. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1110.062291] env[62952]: DEBUG oslo_concurrency.lockutils [req-d590c8c1-282b-4433-a599-89ee6842eb94 req-52776f0b-e393-41a6-b940-637af4b64cba service nova] Acquiring lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.062438] env[62952]: DEBUG oslo_concurrency.lockutils [req-d590c8c1-282b-4433-a599-89ee6842eb94 req-52776f0b-e393-41a6-b940-637af4b64cba service nova] Acquired lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.062604] env[62952]: DEBUG nova.network.neutron [req-d590c8c1-282b-4433-a599-89ee6842eb94 req-52776f0b-e393-41a6-b940-637af4b64cba service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Refreshing network info cache for port ddd93d49-eb57-4af9-a9bb-35d8269b939d {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1110.070041] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5202698a-a639-4f0b-07ce-d5fade8ca9ff, 'name': SearchDatastore_Task, 'duration_secs': 0.010312} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.071315] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-638a5a00-f55b-4634-8a81-335e5edc52cb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.077403] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1110.077403] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52739a75-a047-4dad-40d0-c1ad8c7fd145" [ 1110.077403] env[62952]: _type = "Task" [ 1110.077403] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.085650] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52739a75-a047-4dad-40d0-c1ad8c7fd145, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.376351] env[62952]: DEBUG nova.scheduler.client.report [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1110.477125] env[62952]: DEBUG nova.network.neutron [req-08de22cf-0829-4ffe-b5f5-12e9b4198ecc req-c4207f99-8397-4dfc-957b-9add08d4243d service nova] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updated VIF entry in instance network info cache for port e6fdb64f-1d51-4b10-b297-9f760ae81c95. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1110.477555] env[62952]: DEBUG nova.network.neutron [req-08de22cf-0829-4ffe-b5f5-12e9b4198ecc req-c4207f99-8397-4dfc-957b-9add08d4243d service nova] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating instance_info_cache with network_info: [{"id": "e6fdb64f-1d51-4b10-b297-9f760ae81c95", "address": "fa:16:3e:95:ff:4c", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6fdb64f-1d", "ovs_interfaceid": "e6fdb64f-1d51-4b10-b297-9f760ae81c95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.588572] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52739a75-a047-4dad-40d0-c1ad8c7fd145, 'name': SearchDatastore_Task, 'duration_secs': 0.023225} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.588902] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.589190] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278/11fcbf44-5b9b-4762-8be8-f4e1d0cfa278.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1110.589461] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-48f2bb94-9694-4bc9-990d-b2fab5e9073a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.596905] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1110.596905] env[62952]: value = "task-1263900" [ 1110.596905] env[62952]: _type = "Task" [ 1110.596905] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.604326] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263900, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.815452] env[62952]: DEBUG nova.network.neutron [req-d590c8c1-282b-4433-a599-89ee6842eb94 req-52776f0b-e393-41a6-b940-637af4b64cba service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Updated VIF entry in instance network info cache for port ddd93d49-eb57-4af9-a9bb-35d8269b939d. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1110.816102] env[62952]: DEBUG nova.network.neutron [req-d590c8c1-282b-4433-a599-89ee6842eb94 req-52776f0b-e393-41a6-b940-637af4b64cba service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Updating instance_info_cache with network_info: [{"id": "ddd93d49-eb57-4af9-a9bb-35d8269b939d", "address": "fa:16:3e:e3:8d:93", "network": {"id": "247a88e9-e8e8-46af-ac41-6777f0782fe0", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1297422061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.226", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d93803f89ca64e66b8271e7b0bb48983", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapddd93d49-eb", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.881623] env[62952]: DEBUG oslo_concurrency.lockutils [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.208s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.882336] env[62952]: DEBUG nova.compute.manager [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1110.886426] env[62952]: DEBUG oslo_concurrency.lockutils [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.664s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.888763] env[62952]: INFO nova.compute.claims [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1110.980673] env[62952]: DEBUG oslo_concurrency.lockutils [req-08de22cf-0829-4ffe-b5f5-12e9b4198ecc req-c4207f99-8397-4dfc-957b-9add08d4243d service nova] Releasing lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.110261] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263900, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.318688] env[62952]: DEBUG oslo_concurrency.lockutils [req-d590c8c1-282b-4433-a599-89ee6842eb94 req-52776f0b-e393-41a6-b940-637af4b64cba service nova] Releasing lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.397177] env[62952]: DEBUG nova.compute.utils [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1111.398688] env[62952]: DEBUG nova.compute.manager [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1111.399070] env[62952]: DEBUG nova.network.neutron [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1111.437969] env[62952]: DEBUG nova.policy [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '166cce5924454e73bc7deb1c77191f66', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e2f92c0076c14f15b5a4b8188f50db8e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1111.608323] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263900, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.568252} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.609124] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278/11fcbf44-5b9b-4762-8be8-f4e1d0cfa278.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1111.609124] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1111.609124] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0046bd33-aa94-4ec8-9a42-5fe3db3ef803 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.615970] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1111.615970] env[62952]: value = "task-1263901" [ 1111.615970] env[62952]: _type = "Task" [ 1111.615970] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.623338] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263901, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.701841] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "42793459-89e6-4941-9b20-cbe52a241c6d" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.708881] env[62952]: DEBUG nova.network.neutron [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Successfully created port: 19fcce34-055a-492e-8c73-a39767f93814 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1111.903930] env[62952]: DEBUG nova.compute.manager [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1112.008764] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6fce44a-9d37-42b8-b9a4-e36872de029c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.016562] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6560a7c-a4e7-4130-b488-d445d11c5c7a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.046156] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b47612b7-c265-4865-b6ac-9f3d5d3b24d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.054028] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7fc008-853f-452b-81ac-1d6a70e05768 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.068730] env[62952]: DEBUG nova.compute.provider_tree [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1112.125879] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263901, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065285} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.126161] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1112.126950] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa39ddd-7e8e-4285-a987-812dd58b1b0c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.149206] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278/11fcbf44-5b9b-4762-8be8-f4e1d0cfa278.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1112.149813] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba29ad2c-8313-48b8-9dae-dbdc7d068ac8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.169663] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1112.169663] env[62952]: value = "task-1263902" [ 1112.169663] env[62952]: _type = "Task" [ 1112.169663] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.177475] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263902, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.571981] env[62952]: DEBUG nova.scheduler.client.report [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1112.680321] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263902, 'name': ReconfigVM_Task, 'duration_secs': 0.27273} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.680672] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278/11fcbf44-5b9b-4762-8be8-f4e1d0cfa278.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1112.681477] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9e105198-5096-4990-b477-92ee860a7207 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.687491] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1112.687491] env[62952]: value = "task-1263903" [ 1112.687491] env[62952]: _type = "Task" [ 1112.687491] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.695771] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263903, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.913681] env[62952]: DEBUG nova.compute.manager [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1112.937978] env[62952]: DEBUG nova.virt.hardware [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1112.938367] env[62952]: DEBUG nova.virt.hardware [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1112.938536] env[62952]: DEBUG nova.virt.hardware [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1112.938726] env[62952]: DEBUG nova.virt.hardware [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1112.938878] env[62952]: DEBUG nova.virt.hardware [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1112.939045] env[62952]: DEBUG nova.virt.hardware [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1112.939538] env[62952]: DEBUG nova.virt.hardware [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1112.939538] env[62952]: DEBUG nova.virt.hardware [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1112.939630] env[62952]: DEBUG nova.virt.hardware [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1112.939754] env[62952]: DEBUG nova.virt.hardware [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1112.939930] env[62952]: DEBUG nova.virt.hardware [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1112.940815] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d98704-6626-48cf-8051-56561e4870a2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.949271] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3462bf2-f972-43e2-b943-c4698f093f63 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.080177] env[62952]: DEBUG oslo_concurrency.lockutils [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.191s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.080177] env[62952]: DEBUG nova.compute.manager [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1113.081204] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.822s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.081595] env[62952]: DEBUG nova.objects.instance [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lazy-loading 'resources' on Instance uuid 42793459-89e6-4941-9b20-cbe52a241c6d {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1113.088943] env[62952]: DEBUG nova.compute.manager [req-d34f776d-27ca-48ea-9cd5-b042df84cd72 req-cedadb3f-81a5-4813-be09-9ef1136036c6 service nova] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Received event network-vif-plugged-19fcce34-055a-492e-8c73-a39767f93814 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1113.088943] env[62952]: DEBUG oslo_concurrency.lockutils [req-d34f776d-27ca-48ea-9cd5-b042df84cd72 req-cedadb3f-81a5-4813-be09-9ef1136036c6 service nova] Acquiring lock "f92ef782-fd64-47f0-8145-9261a1c3005c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.088943] env[62952]: DEBUG oslo_concurrency.lockutils [req-d34f776d-27ca-48ea-9cd5-b042df84cd72 req-cedadb3f-81a5-4813-be09-9ef1136036c6 service nova] Lock "f92ef782-fd64-47f0-8145-9261a1c3005c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.088943] env[62952]: DEBUG oslo_concurrency.lockutils [req-d34f776d-27ca-48ea-9cd5-b042df84cd72 req-cedadb3f-81a5-4813-be09-9ef1136036c6 service nova] Lock "f92ef782-fd64-47f0-8145-9261a1c3005c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.088943] env[62952]: DEBUG nova.compute.manager [req-d34f776d-27ca-48ea-9cd5-b042df84cd72 req-cedadb3f-81a5-4813-be09-9ef1136036c6 service nova] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] No waiting events found dispatching network-vif-plugged-19fcce34-055a-492e-8c73-a39767f93814 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1113.088943] env[62952]: WARNING nova.compute.manager [req-d34f776d-27ca-48ea-9cd5-b042df84cd72 req-cedadb3f-81a5-4813-be09-9ef1136036c6 service nova] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Received unexpected event network-vif-plugged-19fcce34-055a-492e-8c73-a39767f93814 for instance with vm_state building and task_state spawning. [ 1113.173855] env[62952]: DEBUG nova.network.neutron [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Successfully updated port: 19fcce34-055a-492e-8c73-a39767f93814 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1113.196932] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263903, 'name': Rename_Task, 'duration_secs': 0.134672} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.197265] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1113.197517] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fdf5a2b0-a83a-46f7-925a-7aa1e72d0c9b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.203621] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1113.203621] env[62952]: value = "task-1263904" [ 1113.203621] env[62952]: _type = "Task" [ 1113.203621] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.211010] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263904, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.585011] env[62952]: DEBUG nova.compute.utils [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1113.586571] env[62952]: DEBUG nova.compute.manager [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1113.586741] env[62952]: DEBUG nova.network.neutron [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1113.588953] env[62952]: DEBUG nova.objects.instance [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lazy-loading 'numa_topology' on Instance uuid 42793459-89e6-4941-9b20-cbe52a241c6d {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1113.628255] env[62952]: DEBUG nova.policy [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '567053f9b1ee48febe32adda91ba0e99', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '641323d095f04c1a8795e6fe6054e338', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1113.677542] env[62952]: DEBUG oslo_concurrency.lockutils [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.677661] env[62952]: DEBUG oslo_concurrency.lockutils [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.677813] env[62952]: DEBUG nova.network.neutron [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1113.713593] env[62952]: DEBUG oslo_vmware.api [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263904, 'name': PowerOnVM_Task, 'duration_secs': 0.431232} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.713865] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1113.714084] env[62952]: INFO nova.compute.manager [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Took 8.70 seconds to spawn the instance on the hypervisor. [ 1113.714274] env[62952]: DEBUG nova.compute.manager [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1113.715079] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b05dbe5e-abb2-4f29-8f66-4df9f8ed1541 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.865455] env[62952]: DEBUG nova.network.neutron [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Successfully created port: 9a78ec8c-90bc-4cf8-9e3a-9bcdd922357a {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1114.091430] env[62952]: DEBUG nova.compute.manager [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1114.094274] env[62952]: DEBUG nova.objects.base [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Object Instance<42793459-89e6-4941-9b20-cbe52a241c6d> lazy-loaded attributes: resources,numa_topology {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1114.211370] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11cab5e9-2d92-45e9-9ccd-d481aa402e6f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.214925] env[62952]: DEBUG nova.network.neutron [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1114.221446] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4623d722-ad7c-4e69-86ec-ae5868c63b7d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.234532] env[62952]: INFO nova.compute.manager [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Took 13.63 seconds to build instance. [ 1114.263350] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fabd565c-d048-4646-9414-20a6ff4ca571 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.673s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.264131] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4098cfed-4fea-4601-98f0-787d64278108 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.272446] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd86d65d-cae7-4c95-8267-fb091ed57345 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.286420] env[62952]: DEBUG nova.compute.provider_tree [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.430963] env[62952]: DEBUG nova.network.neutron [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance_info_cache with network_info: [{"id": "19fcce34-055a-492e-8c73-a39767f93814", "address": "fa:16:3e:71:3d:5f", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19fcce34-05", "ovs_interfaceid": "19fcce34-055a-492e-8c73-a39767f93814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.791123] env[62952]: DEBUG nova.scheduler.client.report [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1114.934252] env[62952]: DEBUG oslo_concurrency.lockutils [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.934597] env[62952]: DEBUG nova.compute.manager [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Instance network_info: |[{"id": "19fcce34-055a-492e-8c73-a39767f93814", "address": "fa:16:3e:71:3d:5f", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19fcce34-05", "ovs_interfaceid": "19fcce34-055a-492e-8c73-a39767f93814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1114.935491] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:3d:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3f499bc9-78da-46c1-9274-19edf26d31cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '19fcce34-055a-492e-8c73-a39767f93814', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1114.943017] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Creating folder: Project (e2f92c0076c14f15b5a4b8188f50db8e). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1114.943374] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fed0e5ba-a5a2-4e0e-b29e-b77bad2fe2a1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.955826] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Created folder: Project (e2f92c0076c14f15b5a4b8188f50db8e) in parent group-v271811. [ 1114.956029] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Creating folder: Instances. Parent ref: group-v271976. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1114.956284] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7b38909d-197d-412e-bf78-ca8560cdeaf6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.966420] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Created folder: Instances in parent group-v271976. [ 1114.966614] env[62952]: DEBUG oslo.service.loopingcall [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1114.966809] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1114.967057] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4bb276af-1385-4987-b1cb-052a6a3ec335 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.985540] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1114.985540] env[62952]: value = "task-1263907" [ 1114.985540] env[62952]: _type = "Task" [ 1114.985540] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.992879] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263907, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.104130] env[62952]: DEBUG nova.compute.manager [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1115.118656] env[62952]: DEBUG nova.compute.manager [req-604e3417-a7fa-42fb-9bcb-c840a436bbdf req-699c7720-0c2e-4935-ab42-4dbf0b4566e7 service nova] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Received event network-changed-19fcce34-055a-492e-8c73-a39767f93814 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1115.118656] env[62952]: DEBUG nova.compute.manager [req-604e3417-a7fa-42fb-9bcb-c840a436bbdf req-699c7720-0c2e-4935-ab42-4dbf0b4566e7 service nova] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Refreshing instance network info cache due to event network-changed-19fcce34-055a-492e-8c73-a39767f93814. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1115.118656] env[62952]: DEBUG oslo_concurrency.lockutils [req-604e3417-a7fa-42fb-9bcb-c840a436bbdf req-699c7720-0c2e-4935-ab42-4dbf0b4566e7 service nova] Acquiring lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.118760] env[62952]: DEBUG oslo_concurrency.lockutils [req-604e3417-a7fa-42fb-9bcb-c840a436bbdf req-699c7720-0c2e-4935-ab42-4dbf0b4566e7 service nova] Acquired lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.119239] env[62952]: DEBUG nova.network.neutron [req-604e3417-a7fa-42fb-9bcb-c840a436bbdf req-699c7720-0c2e-4935-ab42-4dbf0b4566e7 service nova] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Refreshing network info cache for port 19fcce34-055a-492e-8c73-a39767f93814 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1115.134231] env[62952]: DEBUG nova.virt.hardware [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1115.134537] env[62952]: DEBUG nova.virt.hardware [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1115.134703] env[62952]: DEBUG nova.virt.hardware [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1115.134888] env[62952]: DEBUG nova.virt.hardware [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1115.135046] env[62952]: DEBUG nova.virt.hardware [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1115.135214] env[62952]: DEBUG nova.virt.hardware [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1115.135429] env[62952]: DEBUG nova.virt.hardware [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1115.135591] env[62952]: DEBUG nova.virt.hardware [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1115.135761] env[62952]: DEBUG nova.virt.hardware [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1115.135923] env[62952]: DEBUG nova.virt.hardware [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1115.136113] env[62952]: DEBUG nova.virt.hardware [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1115.137517] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e44036c5-da4c-483f-97a3-56e3135c4cb8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.147428] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bdf6053-0ad1-498f-8178-5b85b6dc0cb4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.296334] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.215s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.495637] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263907, 'name': CreateVM_Task, 'duration_secs': 0.36383} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.495816] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1115.496545] env[62952]: DEBUG oslo_concurrency.lockutils [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.496713] env[62952]: DEBUG oslo_concurrency.lockutils [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.497086] env[62952]: DEBUG oslo_concurrency.lockutils [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1115.497347] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c77d199-3dfc-4f42-bb9a-3a40f85321b7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.501609] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1115.501609] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521d7ca8-c8bc-ee64-f13b-7d7eb74cfd6c" [ 1115.501609] env[62952]: _type = "Task" [ 1115.501609] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.508587] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521d7ca8-c8bc-ee64-f13b-7d7eb74cfd6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.535797] env[62952]: DEBUG nova.network.neutron [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Successfully updated port: 9a78ec8c-90bc-4cf8-9e3a-9bcdd922357a {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1115.807041] env[62952]: DEBUG nova.network.neutron [req-604e3417-a7fa-42fb-9bcb-c840a436bbdf req-699c7720-0c2e-4935-ab42-4dbf0b4566e7 service nova] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updated VIF entry in instance network info cache for port 19fcce34-055a-492e-8c73-a39767f93814. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1115.807342] env[62952]: DEBUG nova.network.neutron [req-604e3417-a7fa-42fb-9bcb-c840a436bbdf req-699c7720-0c2e-4935-ab42-4dbf0b4566e7 service nova] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance_info_cache with network_info: [{"id": "19fcce34-055a-492e-8c73-a39767f93814", "address": "fa:16:3e:71:3d:5f", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19fcce34-05", "ovs_interfaceid": "19fcce34-055a-492e-8c73-a39767f93814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.808712] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2bec4bed-7a6f-45bd-a7eb-45c22e56d427 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "42793459-89e6-4941-9b20-cbe52a241c6d" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.785s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.809527] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "42793459-89e6-4941-9b20-cbe52a241c6d" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 4.108s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.809721] env[62952]: INFO nova.compute.manager [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Unshelving [ 1115.843558] env[62952]: DEBUG nova.compute.manager [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Stashing vm_state: active {{(pid=62952) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1116.017152] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]521d7ca8-c8bc-ee64-f13b-7d7eb74cfd6c, 'name': SearchDatastore_Task, 'duration_secs': 0.040938} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.017635] env[62952]: DEBUG oslo_concurrency.lockutils [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.018026] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1116.018404] env[62952]: DEBUG oslo_concurrency.lockutils [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.018671] env[62952]: DEBUG oslo_concurrency.lockutils [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.018986] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1116.019390] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2653093-99ac-4d47-bbaf-06a74249e77c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.030013] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1116.030307] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1116.031398] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1479e1b9-f39d-4c82-adc9-c2bc77eabe53 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.038068] env[62952]: DEBUG oslo_concurrency.lockutils [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Acquiring lock "refresh_cache-148e52e8-07ef-4b3d-a43b-eefc6820b248" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.038160] env[62952]: DEBUG oslo_concurrency.lockutils [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Acquired lock "refresh_cache-148e52e8-07ef-4b3d-a43b-eefc6820b248" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.038382] env[62952]: DEBUG nova.network.neutron [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1116.041193] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1116.041193] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524fff82-e68f-6f06-d69c-cc1f2881206c" [ 1116.041193] env[62952]: _type = "Task" [ 1116.041193] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.049676] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524fff82-e68f-6f06-d69c-cc1f2881206c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.312125] env[62952]: DEBUG oslo_concurrency.lockutils [req-604e3417-a7fa-42fb-9bcb-c840a436bbdf req-699c7720-0c2e-4935-ab42-4dbf0b4566e7 service nova] Releasing lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.365083] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.365207] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.552960] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]524fff82-e68f-6f06-d69c-cc1f2881206c, 'name': SearchDatastore_Task, 'duration_secs': 0.012579} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.553834] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2540345b-8c02-44ed-8379-51dac7a43d8d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.560024] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1116.560024] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5275f043-fbaa-3e92-8e95-aeed4bf911a1" [ 1116.560024] env[62952]: _type = "Task" [ 1116.560024] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.569330] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5275f043-fbaa-3e92-8e95-aeed4bf911a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.575222] env[62952]: DEBUG nova.network.neutron [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1116.715425] env[62952]: DEBUG nova.network.neutron [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Updating instance_info_cache with network_info: [{"id": "9a78ec8c-90bc-4cf8-9e3a-9bcdd922357a", "address": "fa:16:3e:97:7c:bc", "network": {"id": "c3044260-50a6-407c-bb81-3a0ce7ba1a34", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1737582734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "641323d095f04c1a8795e6fe6054e338", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a78ec8c-90", "ovs_interfaceid": "9a78ec8c-90bc-4cf8-9e3a-9bcdd922357a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.835233] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.870992] env[62952]: INFO nova.compute.claims [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1117.071051] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5275f043-fbaa-3e92-8e95-aeed4bf911a1, 'name': SearchDatastore_Task, 'duration_secs': 0.011314} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.071051] env[62952]: DEBUG oslo_concurrency.lockutils [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.071051] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] f92ef782-fd64-47f0-8145-9261a1c3005c/f92ef782-fd64-47f0-8145-9261a1c3005c.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1117.071358] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-10d42f7b-8416-4498-acb8-f83dc891556c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.078281] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1117.078281] env[62952]: value = "task-1263908" [ 1117.078281] env[62952]: _type = "Task" [ 1117.078281] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.085722] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263908, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.144295] env[62952]: DEBUG nova.compute.manager [req-24542dd0-37de-4baf-a340-0995e3a412df req-22dd0fc4-fbde-4dd4-8aef-8ad24aa8456a service nova] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Received event network-vif-plugged-9a78ec8c-90bc-4cf8-9e3a-9bcdd922357a {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1117.144598] env[62952]: DEBUG oslo_concurrency.lockutils [req-24542dd0-37de-4baf-a340-0995e3a412df req-22dd0fc4-fbde-4dd4-8aef-8ad24aa8456a service nova] Acquiring lock "148e52e8-07ef-4b3d-a43b-eefc6820b248-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.144863] env[62952]: DEBUG oslo_concurrency.lockutils [req-24542dd0-37de-4baf-a340-0995e3a412df req-22dd0fc4-fbde-4dd4-8aef-8ad24aa8456a service nova] Lock "148e52e8-07ef-4b3d-a43b-eefc6820b248-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.145080] env[62952]: DEBUG oslo_concurrency.lockutils [req-24542dd0-37de-4baf-a340-0995e3a412df req-22dd0fc4-fbde-4dd4-8aef-8ad24aa8456a service nova] Lock "148e52e8-07ef-4b3d-a43b-eefc6820b248-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.145280] env[62952]: DEBUG nova.compute.manager [req-24542dd0-37de-4baf-a340-0995e3a412df req-22dd0fc4-fbde-4dd4-8aef-8ad24aa8456a service nova] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] No waiting events found dispatching network-vif-plugged-9a78ec8c-90bc-4cf8-9e3a-9bcdd922357a {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1117.145479] env[62952]: WARNING nova.compute.manager [req-24542dd0-37de-4baf-a340-0995e3a412df req-22dd0fc4-fbde-4dd4-8aef-8ad24aa8456a service nova] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Received unexpected event network-vif-plugged-9a78ec8c-90bc-4cf8-9e3a-9bcdd922357a for instance with vm_state building and task_state spawning. [ 1117.145678] env[62952]: DEBUG nova.compute.manager [req-24542dd0-37de-4baf-a340-0995e3a412df req-22dd0fc4-fbde-4dd4-8aef-8ad24aa8456a service nova] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Received event network-changed-9a78ec8c-90bc-4cf8-9e3a-9bcdd922357a {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1117.145841] env[62952]: DEBUG nova.compute.manager [req-24542dd0-37de-4baf-a340-0995e3a412df req-22dd0fc4-fbde-4dd4-8aef-8ad24aa8456a service nova] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Refreshing instance network info cache due to event network-changed-9a78ec8c-90bc-4cf8-9e3a-9bcdd922357a. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1117.146022] env[62952]: DEBUG oslo_concurrency.lockutils [req-24542dd0-37de-4baf-a340-0995e3a412df req-22dd0fc4-fbde-4dd4-8aef-8ad24aa8456a service nova] Acquiring lock "refresh_cache-148e52e8-07ef-4b3d-a43b-eefc6820b248" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1117.218378] env[62952]: DEBUG oslo_concurrency.lockutils [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Releasing lock "refresh_cache-148e52e8-07ef-4b3d-a43b-eefc6820b248" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.218702] env[62952]: DEBUG nova.compute.manager [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Instance network_info: |[{"id": "9a78ec8c-90bc-4cf8-9e3a-9bcdd922357a", "address": "fa:16:3e:97:7c:bc", "network": {"id": "c3044260-50a6-407c-bb81-3a0ce7ba1a34", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1737582734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "641323d095f04c1a8795e6fe6054e338", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a78ec8c-90", "ovs_interfaceid": "9a78ec8c-90bc-4cf8-9e3a-9bcdd922357a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1117.219066] env[62952]: DEBUG oslo_concurrency.lockutils [req-24542dd0-37de-4baf-a340-0995e3a412df req-22dd0fc4-fbde-4dd4-8aef-8ad24aa8456a service nova] Acquired lock "refresh_cache-148e52e8-07ef-4b3d-a43b-eefc6820b248" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.219339] env[62952]: DEBUG nova.network.neutron [req-24542dd0-37de-4baf-a340-0995e3a412df req-22dd0fc4-fbde-4dd4-8aef-8ad24aa8456a service nova] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Refreshing network info cache for port 9a78ec8c-90bc-4cf8-9e3a-9bcdd922357a {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1117.220596] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:97:7c:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '678ebbe4-4c53-4eaf-a689-93981310f37d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9a78ec8c-90bc-4cf8-9e3a-9bcdd922357a', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1117.228048] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Creating folder: Project (641323d095f04c1a8795e6fe6054e338). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1117.229037] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-24107987-0bc3-4761-9a65-44e6f2828f9e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.239787] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Created folder: Project (641323d095f04c1a8795e6fe6054e338) in parent group-v271811. [ 1117.239983] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Creating folder: Instances. Parent ref: group-v271979. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1117.240230] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2894cba1-1c4e-455a-831c-93c28d379bf6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.249022] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Created folder: Instances in parent group-v271979. [ 1117.249265] env[62952]: DEBUG oslo.service.loopingcall [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1117.249456] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1117.249664] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2bdc385f-8207-4457-833d-61edeb57c76d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.269182] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1117.269182] env[62952]: value = "task-1263911" [ 1117.269182] env[62952]: _type = "Task" [ 1117.269182] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.277371] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263911, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.377615] env[62952]: INFO nova.compute.resource_tracker [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Updating resource usage from migration 1cead09a-f727-4baf-84e9-9e025e5ff184 [ 1117.510209] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65fa7476-74ad-4cdc-9d9e-51bf5adaf15f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.519059] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8cde47-9536-4ed5-9b6d-6bce914604d0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.553176] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1776df-1455-4835-ad7c-d84954a92e6d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.561550] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1494b3-ff2f-40ec-a4ae-4d9bf54d0550 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.576783] env[62952]: DEBUG nova.compute.provider_tree [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1117.588494] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263908, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.779947] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263911, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.935134] env[62952]: DEBUG nova.network.neutron [req-24542dd0-37de-4baf-a340-0995e3a412df req-22dd0fc4-fbde-4dd4-8aef-8ad24aa8456a service nova] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Updated VIF entry in instance network info cache for port 9a78ec8c-90bc-4cf8-9e3a-9bcdd922357a. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1117.935673] env[62952]: DEBUG nova.network.neutron [req-24542dd0-37de-4baf-a340-0995e3a412df req-22dd0fc4-fbde-4dd4-8aef-8ad24aa8456a service nova] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Updating instance_info_cache with network_info: [{"id": "9a78ec8c-90bc-4cf8-9e3a-9bcdd922357a", "address": "fa:16:3e:97:7c:bc", "network": {"id": "c3044260-50a6-407c-bb81-3a0ce7ba1a34", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1737582734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "641323d095f04c1a8795e6fe6054e338", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a78ec8c-90", "ovs_interfaceid": "9a78ec8c-90bc-4cf8-9e3a-9bcdd922357a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.079968] env[62952]: DEBUG nova.scheduler.client.report [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1118.092511] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263908, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.660537} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.092789] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] f92ef782-fd64-47f0-8145-9261a1c3005c/f92ef782-fd64-47f0-8145-9261a1c3005c.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1118.093018] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1118.093282] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-226bd65b-537b-4938-b09f-0820499cd607 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.100289] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1118.100289] env[62952]: value = "task-1263912" [ 1118.100289] env[62952]: _type = "Task" [ 1118.100289] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.108311] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263912, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.280145] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263911, 'name': CreateVM_Task, 'duration_secs': 0.571805} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.280332] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1118.281077] env[62952]: DEBUG oslo_concurrency.lockutils [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.281303] env[62952]: DEBUG oslo_concurrency.lockutils [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.281665] env[62952]: DEBUG oslo_concurrency.lockutils [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1118.281953] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-601a17d8-56f0-42b3-b478-3cd843f58a98 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.286127] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Waiting for the task: (returnval){ [ 1118.286127] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]527f6e94-47fc-4630-3b51-12ec1125288b" [ 1118.286127] env[62952]: _type = "Task" [ 1118.286127] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.293215] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]527f6e94-47fc-4630-3b51-12ec1125288b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.437996] env[62952]: DEBUG oslo_concurrency.lockutils [req-24542dd0-37de-4baf-a340-0995e3a412df req-22dd0fc4-fbde-4dd4-8aef-8ad24aa8456a service nova] Releasing lock "refresh_cache-148e52e8-07ef-4b3d-a43b-eefc6820b248" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.588351] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.223s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.588578] env[62952]: INFO nova.compute.manager [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Migrating [ 1118.595022] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.760s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.595650] env[62952]: DEBUG nova.objects.instance [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lazy-loading 'pci_requests' on Instance uuid 42793459-89e6-4941-9b20-cbe52a241c6d {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1118.613402] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263912, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066821} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.614040] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1118.614779] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63af0976-33f9-453f-bcf4-5e1964f6c2cd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.637725] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] f92ef782-fd64-47f0-8145-9261a1c3005c/f92ef782-fd64-47f0-8145-9261a1c3005c.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1118.638181] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e86407e-4d45-44d4-b184-d0039c5d4d4d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.657330] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1118.657330] env[62952]: value = "task-1263913" [ 1118.657330] env[62952]: _type = "Task" [ 1118.657330] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.665353] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263913, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.796193] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]527f6e94-47fc-4630-3b51-12ec1125288b, 'name': SearchDatastore_Task, 'duration_secs': 0.0949} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.796514] env[62952]: DEBUG oslo_concurrency.lockutils [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.796783] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1118.797069] env[62952]: DEBUG oslo_concurrency.lockutils [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.797235] env[62952]: DEBUG oslo_concurrency.lockutils [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.797422] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1118.797678] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f6080580-d6a3-491d-96f3-35907c60cfea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.805344] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1118.805516] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1118.806246] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67bacbf5-1530-42c8-8f56-3579bf154686 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.811130] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Waiting for the task: (returnval){ [ 1118.811130] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520004f8-dbaa-7937-2c43-01ef3053149a" [ 1118.811130] env[62952]: _type = "Task" [ 1118.811130] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.818110] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520004f8-dbaa-7937-2c43-01ef3053149a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.104997] env[62952]: DEBUG nova.objects.instance [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lazy-loading 'numa_topology' on Instance uuid 42793459-89e6-4941-9b20-cbe52a241c6d {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1119.107056] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "refresh_cache-11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1119.107328] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "refresh_cache-11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1119.107559] env[62952]: DEBUG nova.network.neutron [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1119.167971] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1119.168343] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263913, 'name': ReconfigVM_Task, 'duration_secs': 0.289011} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.168616] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1119.168775] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1119.170141] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Reconfigured VM instance instance-00000068 to attach disk [datastore2] f92ef782-fd64-47f0-8145-9261a1c3005c/f92ef782-fd64-47f0-8145-9261a1c3005c.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1119.170951] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8214c8ac-13fa-40ad-a8e6-77bce11ffa3a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.177044] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1119.177044] env[62952]: value = "task-1263914" [ 1119.177044] env[62952]: _type = "Task" [ 1119.177044] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.185116] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263914, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.322752] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]520004f8-dbaa-7937-2c43-01ef3053149a, 'name': SearchDatastore_Task, 'duration_secs': 0.007405} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.322752] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64d8b349-f916-4b30-b221-1406ef3f881c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.327822] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Waiting for the task: (returnval){ [ 1119.327822] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f11c86-6e15-3314-844c-fe5fec0f8142" [ 1119.327822] env[62952]: _type = "Task" [ 1119.327822] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.335824] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f11c86-6e15-3314-844c-fe5fec0f8142, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.607598] env[62952]: INFO nova.compute.claims [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1119.690029] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263914, 'name': Rename_Task, 'duration_secs': 0.136477} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.690029] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1119.690179] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c512b05c-f54f-4c9e-af83-c4728a43d39b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.698096] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1119.698096] env[62952]: value = "task-1263915" [ 1119.698096] env[62952]: _type = "Task" [ 1119.698096] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.705096] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263915, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.807892] env[62952]: DEBUG nova.network.neutron [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Updating instance_info_cache with network_info: [{"id": "9d5ab9ea-3d2a-40d3-996d-96b27d38cff5", "address": "fa:16:3e:4e:7f:ef", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d5ab9ea-3d", "ovs_interfaceid": "9d5ab9ea-3d2a-40d3-996d-96b27d38cff5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.837636] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f11c86-6e15-3314-844c-fe5fec0f8142, 'name': SearchDatastore_Task, 'duration_secs': 0.008887} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.837989] env[62952]: DEBUG oslo_concurrency.lockutils [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.838324] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 148e52e8-07ef-4b3d-a43b-eefc6820b248/148e52e8-07ef-4b3d-a43b-eefc6820b248.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1119.838626] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5f59a2ef-99ce-486a-83d3-240b971d8253 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.845016] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Waiting for the task: (returnval){ [ 1119.845016] env[62952]: value = "task-1263916" [ 1119.845016] env[62952]: _type = "Task" [ 1119.845016] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.852511] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': task-1263916, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.209453] env[62952]: DEBUG oslo_vmware.api [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263915, 'name': PowerOnVM_Task, 'duration_secs': 0.439205} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.209805] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1120.209963] env[62952]: INFO nova.compute.manager [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Took 7.30 seconds to spawn the instance on the hypervisor. [ 1120.210176] env[62952]: DEBUG nova.compute.manager [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1120.211012] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ed96db-53fb-453a-9d21-a6a65d3e6721 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.310986] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "refresh_cache-11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1120.355464] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': task-1263916, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.421575} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.355751] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 148e52e8-07ef-4b3d-a43b-eefc6820b248/148e52e8-07ef-4b3d-a43b-eefc6820b248.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1120.355970] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1120.356551] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0b9f7db4-9b7f-42cf-ba1d-21b996074855 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.362828] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Waiting for the task: (returnval){ [ 1120.362828] env[62952]: value = "task-1263917" [ 1120.362828] env[62952]: _type = "Task" [ 1120.362828] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.371193] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': task-1263917, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.733090] env[62952]: INFO nova.compute.manager [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Took 12.08 seconds to build instance. [ 1120.763688] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e864b893-3a12-41a0-8842-02634e50aff2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.771737] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dae0332-13d3-450d-a5e6-4eeb9f8977d3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.802501] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-379a0024-56e8-4c8a-8b12-8930728b8a7f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.809714] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b75425a8-3f5f-4c98-955c-c1d4fced3491 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.825795] env[62952]: DEBUG nova.compute.provider_tree [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.872015] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': task-1263917, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06615} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.872325] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1120.873101] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75bdf46e-3513-44d9-9afa-d596d6b68410 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.896939] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 148e52e8-07ef-4b3d-a43b-eefc6820b248/148e52e8-07ef-4b3d-a43b-eefc6820b248.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1120.897551] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-344ad3b6-7245-48ac-b3e5-08ed7757a40b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.916612] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Waiting for the task: (returnval){ [ 1120.916612] env[62952]: value = "task-1263918" [ 1120.916612] env[62952]: _type = "Task" [ 1120.916612] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.927020] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': task-1263918, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.138937] env[62952]: DEBUG nova.compute.manager [req-782f28eb-3314-40bb-b0ab-21551678ed37 req-fddd13eb-feca-40e6-ad50-03fce79a36e8 service nova] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Received event network-changed-19fcce34-055a-492e-8c73-a39767f93814 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1121.139190] env[62952]: DEBUG nova.compute.manager [req-782f28eb-3314-40bb-b0ab-21551678ed37 req-fddd13eb-feca-40e6-ad50-03fce79a36e8 service nova] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Refreshing instance network info cache due to event network-changed-19fcce34-055a-492e-8c73-a39767f93814. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1121.139413] env[62952]: DEBUG oslo_concurrency.lockutils [req-782f28eb-3314-40bb-b0ab-21551678ed37 req-fddd13eb-feca-40e6-ad50-03fce79a36e8 service nova] Acquiring lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.139560] env[62952]: DEBUG oslo_concurrency.lockutils [req-782f28eb-3314-40bb-b0ab-21551678ed37 req-fddd13eb-feca-40e6-ad50-03fce79a36e8 service nova] Acquired lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.139726] env[62952]: DEBUG nova.network.neutron [req-782f28eb-3314-40bb-b0ab-21551678ed37 req-fddd13eb-feca-40e6-ad50-03fce79a36e8 service nova] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Refreshing network info cache for port 19fcce34-055a-492e-8c73-a39767f93814 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1121.234757] env[62952]: DEBUG oslo_concurrency.lockutils [None req-910c211f-43ad-4065-9bbf-72c64800b8af tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "f92ef782-fd64-47f0-8145-9261a1c3005c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.593s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.328723] env[62952]: DEBUG nova.scheduler.client.report [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1121.428891] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': task-1263918, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.828021] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db266dbc-85c4-44cb-b0e5-7ad933608745 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.845967] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.251s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.848156] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Updating instance '11fcbf44-5b9b-4762-8be8-f4e1d0cfa278' progress to 0 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1121.852306] env[62952]: DEBUG nova.network.neutron [req-782f28eb-3314-40bb-b0ab-21551678ed37 req-fddd13eb-feca-40e6-ad50-03fce79a36e8 service nova] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updated VIF entry in instance network info cache for port 19fcce34-055a-492e-8c73-a39767f93814. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1121.852513] env[62952]: DEBUG nova.network.neutron [req-782f28eb-3314-40bb-b0ab-21551678ed37 req-fddd13eb-feca-40e6-ad50-03fce79a36e8 service nova] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance_info_cache with network_info: [{"id": "19fcce34-055a-492e-8c73-a39767f93814", "address": "fa:16:3e:71:3d:5f", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19fcce34-05", "ovs_interfaceid": "19fcce34-055a-492e-8c73-a39767f93814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1121.893133] env[62952]: INFO nova.network.neutron [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Updating port ddd93d49-eb57-4af9-a9bb-35d8269b939d with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1121.931025] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': task-1263918, 'name': ReconfigVM_Task, 'duration_secs': 0.788853} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.931025] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 148e52e8-07ef-4b3d-a43b-eefc6820b248/148e52e8-07ef-4b3d-a43b-eefc6820b248.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1121.931025] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7135ea4e-1f3f-43fe-8448-ad6e6952bb3f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.936567] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Waiting for the task: (returnval){ [ 1121.936567] env[62952]: value = "task-1263919" [ 1121.936567] env[62952]: _type = "Task" [ 1121.936567] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.944684] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': task-1263919, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.356130] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1122.356522] env[62952]: DEBUG oslo_concurrency.lockutils [req-782f28eb-3314-40bb-b0ab-21551678ed37 req-fddd13eb-feca-40e6-ad50-03fce79a36e8 service nova] Releasing lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.356982] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-056b0eb9-dffc-4085-90a7-99d9063b1022 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.364196] env[62952]: DEBUG oslo_vmware.api [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1122.364196] env[62952]: value = "task-1263920" [ 1122.364196] env[62952]: _type = "Task" [ 1122.364196] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.371860] env[62952]: DEBUG oslo_vmware.api [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263920, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.446565] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': task-1263919, 'name': Rename_Task, 'duration_secs': 0.155995} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.446882] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1122.447168] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef6bdde7-9ded-4f25-bca8-9046ead546b3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.454420] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Waiting for the task: (returnval){ [ 1122.454420] env[62952]: value = "task-1263921" [ 1122.454420] env[62952]: _type = "Task" [ 1122.454420] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.462627] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': task-1263921, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.877569] env[62952]: DEBUG oslo_vmware.api [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263920, 'name': PowerOffVM_Task, 'duration_secs': 0.188677} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.877961] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1122.878229] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Updating instance '11fcbf44-5b9b-4762-8be8-f4e1d0cfa278' progress to 17 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1122.965317] env[62952]: DEBUG oslo_vmware.api [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': task-1263921, 'name': PowerOnVM_Task, 'duration_secs': 0.475693} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.965653] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1122.965935] env[62952]: INFO nova.compute.manager [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Took 7.86 seconds to spawn the instance on the hypervisor. [ 1122.966230] env[62952]: DEBUG nova.compute.manager [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1122.967137] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2e485e-1707-4fa9-8d88-fa546f8a0bb2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.386048] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1123.386048] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1123.386488] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1123.386488] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1123.386488] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1123.386658] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1123.386925] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1123.387150] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1123.387333] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1123.387503] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1123.387947] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1123.392837] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3d3f52c-d041-4c8b-ba46-08acc93e4a69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.405523] env[62952]: DEBUG nova.compute.manager [req-50a9d199-ebd9-4cef-b8da-6dbe3361e6b9 req-f7bd34e8-5437-4554-b519-8144954beff1 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Received event network-vif-plugged-ddd93d49-eb57-4af9-a9bb-35d8269b939d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1123.405523] env[62952]: DEBUG oslo_concurrency.lockutils [req-50a9d199-ebd9-4cef-b8da-6dbe3361e6b9 req-f7bd34e8-5437-4554-b519-8144954beff1 service nova] Acquiring lock "42793459-89e6-4941-9b20-cbe52a241c6d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.405730] env[62952]: DEBUG oslo_concurrency.lockutils [req-50a9d199-ebd9-4cef-b8da-6dbe3361e6b9 req-f7bd34e8-5437-4554-b519-8144954beff1 service nova] Lock "42793459-89e6-4941-9b20-cbe52a241c6d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.405905] env[62952]: DEBUG oslo_concurrency.lockutils [req-50a9d199-ebd9-4cef-b8da-6dbe3361e6b9 req-f7bd34e8-5437-4554-b519-8144954beff1 service nova] Lock "42793459-89e6-4941-9b20-cbe52a241c6d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.406154] env[62952]: DEBUG nova.compute.manager [req-50a9d199-ebd9-4cef-b8da-6dbe3361e6b9 req-f7bd34e8-5437-4554-b519-8144954beff1 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] No waiting events found dispatching network-vif-plugged-ddd93d49-eb57-4af9-a9bb-35d8269b939d {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1123.406334] env[62952]: WARNING nova.compute.manager [req-50a9d199-ebd9-4cef-b8da-6dbe3361e6b9 req-f7bd34e8-5437-4554-b519-8144954beff1 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Received unexpected event network-vif-plugged-ddd93d49-eb57-4af9-a9bb-35d8269b939d for instance with vm_state shelved_offloaded and task_state spawning. [ 1123.412169] env[62952]: DEBUG oslo_vmware.api [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1123.412169] env[62952]: value = "task-1263922" [ 1123.412169] env[62952]: _type = "Task" [ 1123.412169] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.422174] env[62952]: DEBUG oslo_vmware.api [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263922, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.486019] env[62952]: INFO nova.compute.manager [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Took 14.28 seconds to build instance. [ 1123.499733] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.499972] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.500194] env[62952]: DEBUG nova.network.neutron [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1123.922159] env[62952]: DEBUG oslo_vmware.api [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263922, 'name': ReconfigVM_Task, 'duration_secs': 0.444674} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.922526] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Updating instance '11fcbf44-5b9b-4762-8be8-f4e1d0cfa278' progress to 33 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1123.987471] env[62952]: DEBUG oslo_concurrency.lockutils [None req-601e7462-8397-42c8-a086-e1e52b0fe18c tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Lock "148e52e8-07ef-4b3d-a43b-eefc6820b248" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.795s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.200833] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Didn't find any instances for network info cache update. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1124.201079] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.201249] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.201400] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.201551] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.201695] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.201842] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.201987] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1124.202184] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.376605] env[62952]: DEBUG nova.network.neutron [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Updating instance_info_cache with network_info: [{"id": "ddd93d49-eb57-4af9-a9bb-35d8269b939d", "address": "fa:16:3e:e3:8d:93", "network": {"id": "247a88e9-e8e8-46af-ac41-6777f0782fe0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1297422061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.226", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d93803f89ca64e66b8271e7b0bb48983", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddd93d49-eb", "ovs_interfaceid": "ddd93d49-eb57-4af9-a9bb-35d8269b939d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.429718] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1124.430108] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1124.430108] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1124.430303] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1124.430452] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1124.430600] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1124.430836] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1124.431049] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1124.431232] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1124.431402] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1124.431578] env[62952]: DEBUG nova.virt.hardware [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1124.436984] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Reconfiguring VM instance instance-00000067 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1124.437264] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c24253d6-595a-4c60-9f3b-f48df1ed65bc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.456092] env[62952]: DEBUG oslo_vmware.api [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1124.456092] env[62952]: value = "task-1263923" [ 1124.456092] env[62952]: _type = "Task" [ 1124.456092] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.463933] env[62952]: DEBUG oslo_vmware.api [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263923, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.653788] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Acquiring lock "148e52e8-07ef-4b3d-a43b-eefc6820b248" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.654101] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Lock "148e52e8-07ef-4b3d-a43b-eefc6820b248" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.654268] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Acquiring lock "148e52e8-07ef-4b3d-a43b-eefc6820b248-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.654464] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Lock "148e52e8-07ef-4b3d-a43b-eefc6820b248-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.654776] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Lock "148e52e8-07ef-4b3d-a43b-eefc6820b248-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.656912] env[62952]: INFO nova.compute.manager [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Terminating instance [ 1124.658680] env[62952]: DEBUG nova.compute.manager [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1124.658874] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1124.659766] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b686affb-47b4-4861-acfc-a0b3446c935b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.668605] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1124.668830] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1415389-7ef9-4a8e-9e3e-d60f1f193975 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.674696] env[62952]: DEBUG oslo_vmware.api [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Waiting for the task: (returnval){ [ 1124.674696] env[62952]: value = "task-1263924" [ 1124.674696] env[62952]: _type = "Task" [ 1124.674696] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.682045] env[62952]: DEBUG oslo_vmware.api [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': task-1263924, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.705304] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.705548] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.705732] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.705888] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1124.706892] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69441ef-98a4-461f-8b38-5c54e8b1a96a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.714372] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8261d7e-1aae-44a4-b8aa-7b0e30de4b87 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.728343] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a5867c-a935-4c16-adfe-554f75bc460b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.734821] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20a0781-5f56-45d6-9e33-fd337a4abc86 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.763654] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180891MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1124.763785] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.763959] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.879395] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Releasing lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.915176] env[62952]: DEBUG nova.virt.hardware [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='e0b55735932ed856fdc3c50983ac8473',container_format='bare',created_at=2024-10-10T21:39:04Z,direct_url=,disk_format='vmdk',id=79448c63-26a7-44be-ba83-4d295285cf80,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-253857404-shelved',owner='d93803f89ca64e66b8271e7b0bb48983',properties=ImageMetaProps,protected=,size=31664128,status='active',tags=,updated_at=2024-10-10T21:39:17Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1124.915515] env[62952]: DEBUG nova.virt.hardware [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1124.915606] env[62952]: DEBUG nova.virt.hardware [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1124.915789] env[62952]: DEBUG nova.virt.hardware [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1124.915941] env[62952]: DEBUG nova.virt.hardware [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1124.916102] env[62952]: DEBUG nova.virt.hardware [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1124.916325] env[62952]: DEBUG nova.virt.hardware [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1124.916491] env[62952]: DEBUG nova.virt.hardware [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1124.916660] env[62952]: DEBUG nova.virt.hardware [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1124.916868] env[62952]: DEBUG nova.virt.hardware [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1124.917078] env[62952]: DEBUG nova.virt.hardware [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1124.917990] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-087f6b27-4175-4217-8d6d-97e75cfc9c2c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.925485] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6c5664d-50fc-4430-9df5-229e57a6409f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.938385] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:8d:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4fcde7-8926-402a-a9b7-4878d2bc1cf6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ddd93d49-eb57-4af9-a9bb-35d8269b939d', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1124.945567] env[62952]: DEBUG oslo.service.loopingcall [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1124.946116] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1124.946342] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6a14bfd2-b32d-4a4d-a7e4-3a5c6ea87477 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.968559] env[62952]: DEBUG oslo_vmware.api [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263923, 'name': ReconfigVM_Task, 'duration_secs': 0.161863} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.969668] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Reconfigured VM instance instance-00000067 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1124.969928] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1124.969928] env[62952]: value = "task-1263925" [ 1124.969928] env[62952]: _type = "Task" [ 1124.969928] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.970579] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba42419e-a471-49df-89f5-818fed48ed17 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.994237] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278/11fcbf44-5b9b-4762-8be8-f4e1d0cfa278.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1124.996740] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac892fb3-4480-4ab5-9212-09ca42259923 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.009288] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263925, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.014219] env[62952]: DEBUG oslo_vmware.api [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1125.014219] env[62952]: value = "task-1263926" [ 1125.014219] env[62952]: _type = "Task" [ 1125.014219] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.025331] env[62952]: DEBUG oslo_vmware.api [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263926, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.184470] env[62952]: DEBUG oslo_vmware.api [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': task-1263924, 'name': PowerOffVM_Task, 'duration_secs': 0.145167} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.184756] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1125.184943] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1125.185213] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f86a2652-b221-4963-b129-8a755fd6482b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.286487] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1125.286795] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1125.286994] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Deleting the datastore file [datastore2] 148e52e8-07ef-4b3d-a43b-eefc6820b248 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1125.287299] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b08ebeb3-b04c-40f0-8b1f-0ac465415d1f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.296856] env[62952]: DEBUG oslo_vmware.api [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Waiting for the task: (returnval){ [ 1125.296856] env[62952]: value = "task-1263928" [ 1125.296856] env[62952]: _type = "Task" [ 1125.296856] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.305122] env[62952]: DEBUG oslo_vmware.api [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': task-1263928, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.483224] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263925, 'name': CreateVM_Task, 'duration_secs': 0.293266} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.483224] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1125.483224] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/79448c63-26a7-44be-ba83-4d295285cf80" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.483224] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired lock "[datastore2] devstack-image-cache_base/79448c63-26a7-44be-ba83-4d295285cf80" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.483224] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/79448c63-26a7-44be-ba83-4d295285cf80" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1125.483611] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87b4140b-1826-4b66-8b55-980fe640f917 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.487485] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1125.487485] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52545d52-79c7-3167-df33-701f2a3da7f9" [ 1125.487485] env[62952]: _type = "Task" [ 1125.487485] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.494720] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52545d52-79c7-3167-df33-701f2a3da7f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.500489] env[62952]: DEBUG nova.compute.manager [req-471c351c-5c0b-4062-82a5-f2dde3887484 req-5d9977ee-1cc5-46ac-ac41-3a06fc2fdf63 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Received event network-changed-ddd93d49-eb57-4af9-a9bb-35d8269b939d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1125.500693] env[62952]: DEBUG nova.compute.manager [req-471c351c-5c0b-4062-82a5-f2dde3887484 req-5d9977ee-1cc5-46ac-ac41-3a06fc2fdf63 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Refreshing instance network info cache due to event network-changed-ddd93d49-eb57-4af9-a9bb-35d8269b939d. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1125.500916] env[62952]: DEBUG oslo_concurrency.lockutils [req-471c351c-5c0b-4062-82a5-f2dde3887484 req-5d9977ee-1cc5-46ac-ac41-3a06fc2fdf63 service nova] Acquiring lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.501114] env[62952]: DEBUG oslo_concurrency.lockutils [req-471c351c-5c0b-4062-82a5-f2dde3887484 req-5d9977ee-1cc5-46ac-ac41-3a06fc2fdf63 service nova] Acquired lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.501292] env[62952]: DEBUG nova.network.neutron [req-471c351c-5c0b-4062-82a5-f2dde3887484 req-5d9977ee-1cc5-46ac-ac41-3a06fc2fdf63 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Refreshing network info cache for port ddd93d49-eb57-4af9-a9bb-35d8269b939d {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1125.522977] env[62952]: DEBUG oslo_vmware.api [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263926, 'name': ReconfigVM_Task, 'duration_secs': 0.285865} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.523241] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278/11fcbf44-5b9b-4762-8be8-f4e1d0cfa278.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1125.523505] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Updating instance '11fcbf44-5b9b-4762-8be8-f4e1d0cfa278' progress to 50 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1125.775433] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Applying migration context for instance 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278 as it has an incoming, in-progress migration 1cead09a-f727-4baf-84e9-9e025e5ff184. Migration status is migrating {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1125.776586] env[62952]: INFO nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Updating resource usage from migration 1cead09a-f727-4baf-84e9-9e025e5ff184 [ 1125.796393] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance dbca4369-1ed3-493a-9847-9fa1e4293475 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1125.796537] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 77fe79de-8da8-486e-af63-91f9d8196ca9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1125.796689] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 1941365e-d562-4d99-89eb-226a07e52071 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1125.796808] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 43d79d26-4350-4329-bffc-cedfada40ee0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1125.796926] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance f92ef782-fd64-47f0-8145-9261a1c3005c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1125.797052] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 148e52e8-07ef-4b3d-a43b-eefc6820b248 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1125.797168] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 42793459-89e6-4941-9b20-cbe52a241c6d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1125.797281] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Migration 1cead09a-f727-4baf-84e9-9e025e5ff184 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1125.797391] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1125.797569] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1125.797713] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2304MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1125.808422] env[62952]: DEBUG oslo_vmware.api [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Task: {'id': task-1263928, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121997} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.808666] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1125.808852] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1125.809038] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1125.809217] env[62952]: INFO nova.compute.manager [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1125.809448] env[62952]: DEBUG oslo.service.loopingcall [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1125.809634] env[62952]: DEBUG nova.compute.manager [-] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1125.809806] env[62952]: DEBUG nova.network.neutron [-] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1125.896296] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ae314c-3612-4d96-bef5-971234ae536a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.903660] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84caf2d2-a433-49ca-a68a-6ac3d4e33725 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.934084] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a004c52-47cc-4062-9f0f-2ea641f9a147 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.940986] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc69153-cd58-4fb7-894c-a114693eb70a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.953824] env[62952]: DEBUG nova.compute.provider_tree [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1125.997483] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Releasing lock "[datastore2] devstack-image-cache_base/79448c63-26a7-44be-ba83-4d295285cf80" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.998022] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Processing image 79448c63-26a7-44be-ba83-4d295285cf80 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1125.998022] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/79448c63-26a7-44be-ba83-4d295285cf80/79448c63-26a7-44be-ba83-4d295285cf80.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.998164] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquired lock "[datastore2] devstack-image-cache_base/79448c63-26a7-44be-ba83-4d295285cf80/79448c63-26a7-44be-ba83-4d295285cf80.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.998328] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1125.998582] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9fec9237-2eed-4a4a-be35-ce3aafed73fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.008334] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1126.008524] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1126.009712] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1871a80b-7f49-429d-9ae9-5f17542e0808 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.015083] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1126.015083] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f96134-060b-e63b-e594-6c36bb6dde9c" [ 1126.015083] env[62952]: _type = "Task" [ 1126.015083] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.023906] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f96134-060b-e63b-e594-6c36bb6dde9c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.034068] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea920426-aff9-47d2-8687-868ef3684bce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.049406] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f94a26b-4108-4e4c-bab6-87b91dbf20c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.068029] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Updating instance '11fcbf44-5b9b-4762-8be8-f4e1d0cfa278' progress to 67 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1126.225364] env[62952]: DEBUG nova.network.neutron [req-471c351c-5c0b-4062-82a5-f2dde3887484 req-5d9977ee-1cc5-46ac-ac41-3a06fc2fdf63 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Updated VIF entry in instance network info cache for port ddd93d49-eb57-4af9-a9bb-35d8269b939d. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1126.225736] env[62952]: DEBUG nova.network.neutron [req-471c351c-5c0b-4062-82a5-f2dde3887484 req-5d9977ee-1cc5-46ac-ac41-3a06fc2fdf63 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Updating instance_info_cache with network_info: [{"id": "ddd93d49-eb57-4af9-a9bb-35d8269b939d", "address": "fa:16:3e:e3:8d:93", "network": {"id": "247a88e9-e8e8-46af-ac41-6777f0782fe0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1297422061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.226", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d93803f89ca64e66b8271e7b0bb48983", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddd93d49-eb", "ovs_interfaceid": "ddd93d49-eb57-4af9-a9bb-35d8269b939d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.457598] env[62952]: DEBUG nova.scheduler.client.report [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1126.510151] env[62952]: DEBUG nova.network.neutron [-] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.525615] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Preparing fetch location {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1126.525897] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Fetch image to [datastore2] OSTACK_IMG_93295dc4-a6d7-47fb-ac7f-21d46fbba6bd/OSTACK_IMG_93295dc4-a6d7-47fb-ac7f-21d46fbba6bd.vmdk {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1126.526097] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Downloading stream optimized image 79448c63-26a7-44be-ba83-4d295285cf80 to [datastore2] OSTACK_IMG_93295dc4-a6d7-47fb-ac7f-21d46fbba6bd/OSTACK_IMG_93295dc4-a6d7-47fb-ac7f-21d46fbba6bd.vmdk on the data store datastore2 as vApp {{(pid=62952) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1126.526275] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Downloading image file data 79448c63-26a7-44be-ba83-4d295285cf80 to the ESX as VM named 'OSTACK_IMG_93295dc4-a6d7-47fb-ac7f-21d46fbba6bd' {{(pid=62952) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1126.608054] env[62952]: DEBUG oslo_vmware.rw_handles [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1126.608054] env[62952]: value = "resgroup-9" [ 1126.608054] env[62952]: _type = "ResourcePool" [ 1126.608054] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1126.608054] env[62952]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-de123481-7dca-44c4-9432-36b2ce98a2e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.622560] env[62952]: DEBUG nova.network.neutron [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Port 9d5ab9ea-3d2a-40d3-996d-96b27d38cff5 binding to destination host cpu-1 is already ACTIVE {{(pid=62952) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1126.629608] env[62952]: DEBUG oslo_vmware.rw_handles [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lease: (returnval){ [ 1126.629608] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a59fde-5d9d-67b2-0ffe-580039e00bd3" [ 1126.629608] env[62952]: _type = "HttpNfcLease" [ 1126.629608] env[62952]: } obtained for vApp import into resource pool (val){ [ 1126.629608] env[62952]: value = "resgroup-9" [ 1126.629608] env[62952]: _type = "ResourcePool" [ 1126.629608] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1126.630107] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the lease: (returnval){ [ 1126.630107] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a59fde-5d9d-67b2-0ffe-580039e00bd3" [ 1126.630107] env[62952]: _type = "HttpNfcLease" [ 1126.630107] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1126.637680] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1126.637680] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a59fde-5d9d-67b2-0ffe-580039e00bd3" [ 1126.637680] env[62952]: _type = "HttpNfcLease" [ 1126.637680] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1126.729063] env[62952]: DEBUG oslo_concurrency.lockutils [req-471c351c-5c0b-4062-82a5-f2dde3887484 req-5d9977ee-1cc5-46ac-ac41-3a06fc2fdf63 service nova] Releasing lock "refresh_cache-42793459-89e6-4941-9b20-cbe52a241c6d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.962788] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1126.963125] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.199s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.013128] env[62952]: INFO nova.compute.manager [-] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Took 1.20 seconds to deallocate network for instance. [ 1127.138426] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1127.138426] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a59fde-5d9d-67b2-0ffe-580039e00bd3" [ 1127.138426] env[62952]: _type = "HttpNfcLease" [ 1127.138426] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1127.138705] env[62952]: DEBUG oslo_vmware.rw_handles [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1127.138705] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a59fde-5d9d-67b2-0ffe-580039e00bd3" [ 1127.138705] env[62952]: _type = "HttpNfcLease" [ 1127.138705] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1127.139410] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4dd9d0b-6a9c-4cec-ad11-d488916d3851 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.146374] env[62952]: DEBUG oslo_vmware.rw_handles [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e62829-cd53-ba80-0101-932c62e94d30/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1127.146501] env[62952]: DEBUG oslo_vmware.rw_handles [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Creating HTTP connection to write to file with size = 31664128 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e62829-cd53-ba80-0101-932c62e94d30/disk-0.vmdk. {{(pid=62952) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1127.205542] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "1941365e-d562-4d99-89eb-226a07e52071" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.206461] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "1941365e-d562-4d99-89eb-226a07e52071" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.213242] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5c36ea08-baf4-43c6-a6fa-d6a7bfe42b91 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.520506] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.520811] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.521101] env[62952]: DEBUG nova.objects.instance [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Lazy-loading 'resources' on Instance uuid 148e52e8-07ef-4b3d-a43b-eefc6820b248 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1127.527786] env[62952]: DEBUG nova.compute.manager [req-2ff04a63-a0e2-4a2a-9e60-d0fe8864ec39 req-b5353274-0c3a-4c67-b5e5-93eb88e597cb service nova] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Received event network-vif-deleted-9a78ec8c-90bc-4cf8-9e3a-9bcdd922357a {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1127.645236] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.645510] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.645692] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.709504] env[62952]: INFO nova.compute.manager [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Detaching volume 7ee58197-e579-4b82-9820-78f9798fc80e [ 1127.753641] env[62952]: INFO nova.virt.block_device [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Attempting to driver detach volume 7ee58197-e579-4b82-9820-78f9798fc80e from mountpoint /dev/sdb [ 1127.753641] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1127.753793] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271968', 'volume_id': '7ee58197-e579-4b82-9820-78f9798fc80e', 'name': 'volume-7ee58197-e579-4b82-9820-78f9798fc80e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1941365e-d562-4d99-89eb-226a07e52071', 'attached_at': '', 'detached_at': '', 'volume_id': '7ee58197-e579-4b82-9820-78f9798fc80e', 'serial': '7ee58197-e579-4b82-9820-78f9798fc80e'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1127.754889] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9654bbea-41a4-4b6e-9283-57bf205e2c0f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.787924] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc46ef0-cd7a-4cdc-b7d7-62ba5c2e0919 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.798074] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3fca9c1-cae2-4a93-a5a4-184b5405b697 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.826369] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99267c5d-0f14-43bd-93ab-890969032e5e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.841779] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] The volume has not been displaced from its original location: [datastore2] volume-7ee58197-e579-4b82-9820-78f9798fc80e/volume-7ee58197-e579-4b82-9820-78f9798fc80e.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1127.847228] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Reconfiguring VM instance instance-00000059 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1127.852765] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-334e6153-df5e-4dd9-b1c3-fbfee3417ab0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.874856] env[62952]: DEBUG oslo_vmware.api [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1127.874856] env[62952]: value = "task-1263930" [ 1127.874856] env[62952]: _type = "Task" [ 1127.874856] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.884280] env[62952]: DEBUG oslo_vmware.api [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263930, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.187454] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-939772e1-33fb-4f48-b1e8-3a5d33055358 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.198702] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e1f82d8-bf93-48db-9130-c6830013808a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.234913] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc8c049-3262-4d97-8e9a-0ae7f47ae0bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.242692] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b97655-2bb1-44e7-a467-e32803cb0ced {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.256287] env[62952]: DEBUG nova.compute.provider_tree [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1128.315020] env[62952]: DEBUG oslo_vmware.rw_handles [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Completed reading data from the image iterator. {{(pid=62952) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1128.315233] env[62952]: DEBUG oslo_vmware.rw_handles [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e62829-cd53-ba80-0101-932c62e94d30/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1128.316184] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3026fe04-5da8-46ce-9d25-3ed8018195ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.323035] env[62952]: DEBUG oslo_vmware.rw_handles [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e62829-cd53-ba80-0101-932c62e94d30/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1128.323253] env[62952]: DEBUG oslo_vmware.rw_handles [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e62829-cd53-ba80-0101-932c62e94d30/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1128.323483] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-8dc645ed-0991-4d7c-bc18-21254bf1e060 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.386184] env[62952]: DEBUG oslo_vmware.api [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263930, 'name': ReconfigVM_Task, 'duration_secs': 0.264163} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.386449] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Reconfigured VM instance instance-00000059 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1128.391123] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64a96a9f-4f58-4a1c-be11-e02f33021b3b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.406296] env[62952]: DEBUG oslo_vmware.api [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1128.406296] env[62952]: value = "task-1263931" [ 1128.406296] env[62952]: _type = "Task" [ 1128.406296] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.413908] env[62952]: DEBUG oslo_vmware.api [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263931, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.503227] env[62952]: DEBUG oslo_vmware.rw_handles [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e62829-cd53-ba80-0101-932c62e94d30/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1128.503447] env[62952]: INFO nova.virt.vmwareapi.images [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Downloaded image file data 79448c63-26a7-44be-ba83-4d295285cf80 [ 1128.504318] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe90b4d-785c-4f0e-b904-334c82f7856a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.520857] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8c7b72d-58fd-4d60-8c6d-7ac88a9be107 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.554913] env[62952]: INFO nova.virt.vmwareapi.images [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] The imported VM was unregistered [ 1128.557397] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Caching image {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1128.557656] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Creating directory with path [datastore2] devstack-image-cache_base/79448c63-26a7-44be-ba83-4d295285cf80 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1128.557975] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aa84f9f6-cac0-4e23-a458-2a77edbe45df {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.566343] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Created directory with path [datastore2] devstack-image-cache_base/79448c63-26a7-44be-ba83-4d295285cf80 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1128.566520] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_93295dc4-a6d7-47fb-ac7f-21d46fbba6bd/OSTACK_IMG_93295dc4-a6d7-47fb-ac7f-21d46fbba6bd.vmdk to [datastore2] devstack-image-cache_base/79448c63-26a7-44be-ba83-4d295285cf80/79448c63-26a7-44be-ba83-4d295285cf80.vmdk. {{(pid=62952) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1128.566827] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-892e1e89-15c0-4c3d-9806-851a14487eb3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.572420] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1128.572420] env[62952]: value = "task-1263933" [ 1128.572420] env[62952]: _type = "Task" [ 1128.572420] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.579573] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263933, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.681392] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "refresh_cache-11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.681593] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "refresh_cache-11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.681777] env[62952]: DEBUG nova.network.neutron [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1128.759132] env[62952]: DEBUG nova.scheduler.client.report [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1128.915778] env[62952]: DEBUG oslo_vmware.api [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263931, 'name': ReconfigVM_Task, 'duration_secs': 0.148731} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.916125] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271968', 'volume_id': '7ee58197-e579-4b82-9820-78f9798fc80e', 'name': 'volume-7ee58197-e579-4b82-9820-78f9798fc80e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1941365e-d562-4d99-89eb-226a07e52071', 'attached_at': '', 'detached_at': '', 'volume_id': '7ee58197-e579-4b82-9820-78f9798fc80e', 'serial': '7ee58197-e579-4b82-9820-78f9798fc80e'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1129.082886] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263933, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.266951] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.746s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.287736] env[62952]: INFO nova.scheduler.client.report [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Deleted allocations for instance 148e52e8-07ef-4b3d-a43b-eefc6820b248 [ 1129.413969] env[62952]: DEBUG nova.network.neutron [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Updating instance_info_cache with network_info: [{"id": "9d5ab9ea-3d2a-40d3-996d-96b27d38cff5", "address": "fa:16:3e:4e:7f:ef", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d5ab9ea-3d", "ovs_interfaceid": "9d5ab9ea-3d2a-40d3-996d-96b27d38cff5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.460682] env[62952]: DEBUG nova.objects.instance [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lazy-loading 'flavor' on Instance uuid 1941365e-d562-4d99-89eb-226a07e52071 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.584128] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263933, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.795728] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4ae49f2-5645-43a1-a71c-3aa5d40f7c90 tempest-ServerMetadataTestJSON-1585242720 tempest-ServerMetadataTestJSON-1585242720-project-member] Lock "148e52e8-07ef-4b3d-a43b-eefc6820b248" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.141s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.916979] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "refresh_cache-11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.087208] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263933, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.443518] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37234445-9b3d-4318-b700-c7ec67cf5a4f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.465049] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4196ed92-5894-4a70-86da-665f3da239f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.469022] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9d470cce-9d2a-4b72-baa5-2c85e3ec1a3e tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "1941365e-d562-4d99-89eb-226a07e52071" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.263s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.472630] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Updating instance '11fcbf44-5b9b-4762-8be8-f4e1d0cfa278' progress to 83 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1130.585458] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263933, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.981570] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1130.981570] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ab0d244-09b4-4d6d-b741-2c2a4ea0a9f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.987798] env[62952]: DEBUG oslo_vmware.api [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1130.987798] env[62952]: value = "task-1263934" [ 1130.987798] env[62952]: _type = "Task" [ 1130.987798] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.996143] env[62952]: DEBUG oslo_vmware.api [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263934, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.085563] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263933, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.196294} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.085876] env[62952]: INFO nova.virt.vmwareapi.ds_util [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_93295dc4-a6d7-47fb-ac7f-21d46fbba6bd/OSTACK_IMG_93295dc4-a6d7-47fb-ac7f-21d46fbba6bd.vmdk to [datastore2] devstack-image-cache_base/79448c63-26a7-44be-ba83-4d295285cf80/79448c63-26a7-44be-ba83-4d295285cf80.vmdk. [ 1131.086099] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Cleaning up location [datastore2] OSTACK_IMG_93295dc4-a6d7-47fb-ac7f-21d46fbba6bd {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1131.086275] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_93295dc4-a6d7-47fb-ac7f-21d46fbba6bd {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1131.086529] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-800fd9d1-c28c-4e35-bfe8-e4a5379104dd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.092888] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1131.092888] env[62952]: value = "task-1263935" [ 1131.092888] env[62952]: _type = "Task" [ 1131.092888] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.101612] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263935, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.498246] env[62952]: DEBUG oslo_vmware.api [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263934, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.516983] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "1941365e-d562-4d99-89eb-226a07e52071" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.517303] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "1941365e-d562-4d99-89eb-226a07e52071" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.517519] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "1941365e-d562-4d99-89eb-226a07e52071-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.517746] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "1941365e-d562-4d99-89eb-226a07e52071-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.517929] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "1941365e-d562-4d99-89eb-226a07e52071-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.520118] env[62952]: INFO nova.compute.manager [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Terminating instance [ 1131.521859] env[62952]: DEBUG nova.compute.manager [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1131.522090] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1131.523020] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8ab08e-ff83-4d73-b0e7-1dd70b892ae0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.530157] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1131.530373] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-022a0208-a94a-454b-915a-98b08f3f353c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.535841] env[62952]: DEBUG oslo_vmware.api [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1131.535841] env[62952]: value = "task-1263936" [ 1131.535841] env[62952]: _type = "Task" [ 1131.535841] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.544331] env[62952]: DEBUG oslo_vmware.api [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263936, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.604032] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263935, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.042189} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.604032] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1131.604032] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Releasing lock "[datastore2] devstack-image-cache_base/79448c63-26a7-44be-ba83-4d295285cf80/79448c63-26a7-44be-ba83-4d295285cf80.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.604032] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/79448c63-26a7-44be-ba83-4d295285cf80/79448c63-26a7-44be-ba83-4d295285cf80.vmdk to [datastore2] 42793459-89e6-4941-9b20-cbe52a241c6d/42793459-89e6-4941-9b20-cbe52a241c6d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1131.604500] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-75ead996-547f-4581-840e-388b2362340d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.611757] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1131.611757] env[62952]: value = "task-1263937" [ 1131.611757] env[62952]: _type = "Task" [ 1131.611757] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.619194] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263937, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.998090] env[62952]: DEBUG oslo_vmware.api [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263934, 'name': PowerOnVM_Task, 'duration_secs': 0.544348} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.998396] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1131.998591] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ae91432f-51c9-4177-9bb6-a9e9c229e395 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Updating instance '11fcbf44-5b9b-4762-8be8-f4e1d0cfa278' progress to 100 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1132.047556] env[62952]: DEBUG oslo_vmware.api [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263936, 'name': PowerOffVM_Task, 'duration_secs': 0.402573} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.047865] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1132.048051] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1132.048318] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b682d85f-60c4-44d1-854d-950952df66d4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.121898] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263937, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.139749] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1132.139997] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1132.140180] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Deleting the datastore file [datastore2] 1941365e-d562-4d99-89eb-226a07e52071 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1132.140456] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4027508b-a943-42ef-9afd-8f39694df545 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.146937] env[62952]: DEBUG oslo_vmware.api [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1132.146937] env[62952]: value = "task-1263939" [ 1132.146937] env[62952]: _type = "Task" [ 1132.146937] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.155152] env[62952]: DEBUG oslo_vmware.api [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263939, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.623021] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263937, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.657475] env[62952]: DEBUG oslo_vmware.api [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263939, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.125263] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263937, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.157389] env[62952]: DEBUG oslo_vmware.api [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263939, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.626244] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263937, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.657702] env[62952]: DEBUG oslo_vmware.api [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263939, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.126249] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263937, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.158184] env[62952]: DEBUG oslo_vmware.api [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263939, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.897436} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.158422] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1134.158608] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1134.158788] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1134.158965] env[62952]: INFO nova.compute.manager [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Took 2.64 seconds to destroy the instance on the hypervisor. [ 1134.159234] env[62952]: DEBUG oslo.service.loopingcall [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1134.159433] env[62952]: DEBUG nova.compute.manager [-] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1134.159530] env[62952]: DEBUG nova.network.neutron [-] [instance: 1941365e-d562-4d99-89eb-226a07e52071] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1134.189027] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.189252] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.189450] env[62952]: DEBUG nova.compute.manager [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Going to confirm migration 4 {{(pid=62952) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1134.605433] env[62952]: DEBUG nova.compute.manager [req-cae4ed8c-30ec-4fca-a749-a8d403f394e3 req-f56317c8-ba92-474d-817a-8b59ae0637a1 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Received event network-vif-deleted-e1c235af-9501-47c0-a83a-ba40ea7fa8c1 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1134.605551] env[62952]: INFO nova.compute.manager [req-cae4ed8c-30ec-4fca-a749-a8d403f394e3 req-f56317c8-ba92-474d-817a-8b59ae0637a1 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Neutron deleted interface e1c235af-9501-47c0-a83a-ba40ea7fa8c1; detaching it from the instance and deleting it from the info cache [ 1134.605653] env[62952]: DEBUG nova.network.neutron [req-cae4ed8c-30ec-4fca-a749-a8d403f394e3 req-f56317c8-ba92-474d-817a-8b59ae0637a1 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.626732] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263937, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.771767] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "refresh_cache-11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1134.771968] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquired lock "refresh_cache-11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.772172] env[62952]: DEBUG nova.network.neutron [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1134.772370] env[62952]: DEBUG nova.objects.instance [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lazy-loading 'info_cache' on Instance uuid 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1135.083166] env[62952]: DEBUG nova.network.neutron [-] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.109742] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4225c397-7444-4121-b776-af047610581e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.119757] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb97ca86-5381-440e-a3db-c27cd792b61b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.140479] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263937, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.149020] env[62952]: DEBUG nova.compute.manager [req-cae4ed8c-30ec-4fca-a749-a8d403f394e3 req-f56317c8-ba92-474d-817a-8b59ae0637a1 service nova] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Detach interface failed, port_id=e1c235af-9501-47c0-a83a-ba40ea7fa8c1, reason: Instance 1941365e-d562-4d99-89eb-226a07e52071 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1135.587109] env[62952]: INFO nova.compute.manager [-] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Took 1.43 seconds to deallocate network for instance. [ 1135.640522] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263937, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.578951} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.640867] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/79448c63-26a7-44be-ba83-4d295285cf80/79448c63-26a7-44be-ba83-4d295285cf80.vmdk to [datastore2] 42793459-89e6-4941-9b20-cbe52a241c6d/42793459-89e6-4941-9b20-cbe52a241c6d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1135.641568] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f1a18f-9fa3-4134-a556-a234fd9aa355 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.663445] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 42793459-89e6-4941-9b20-cbe52a241c6d/42793459-89e6-4941-9b20-cbe52a241c6d.vmdk or device None with type streamOptimized {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1135.663950] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-621fbdd2-7d24-4738-acf9-6a141032e2f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.683065] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1135.683065] env[62952]: value = "task-1263940" [ 1135.683065] env[62952]: _type = "Task" [ 1135.683065] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.692017] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263940, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.972074] env[62952]: DEBUG nova.network.neutron [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Updating instance_info_cache with network_info: [{"id": "9d5ab9ea-3d2a-40d3-996d-96b27d38cff5", "address": "fa:16:3e:4e:7f:ef", "network": {"id": "5d4150ac-7578-43a8-b9f9-c64894c49a19", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-890859607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff928791a3cf496cb8db2c2a6c47a6a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9f430b-e6f7-4a47-abd0-3cc7bef3e97c", "external-id": "nsx-vlan-transportzone-977", "segmentation_id": 977, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d5ab9ea-3d", "ovs_interfaceid": "9d5ab9ea-3d2a-40d3-996d-96b27d38cff5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.093837] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.094092] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.094326] env[62952]: DEBUG nova.objects.instance [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lazy-loading 'resources' on Instance uuid 1941365e-d562-4d99-89eb-226a07e52071 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1136.192667] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263940, 'name': ReconfigVM_Task, 'duration_secs': 0.367528} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.192944] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 42793459-89e6-4941-9b20-cbe52a241c6d/42793459-89e6-4941-9b20-cbe52a241c6d.vmdk or device None with type streamOptimized {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1136.193596] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f5184953-983c-4c26-a7c4-1a9c14a86186 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.200174] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1136.200174] env[62952]: value = "task-1263941" [ 1136.200174] env[62952]: _type = "Task" [ 1136.200174] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.208135] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263941, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.475478] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Releasing lock "refresh_cache-11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.475804] env[62952]: DEBUG nova.objects.instance [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lazy-loading 'migration_context' on Instance uuid 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1136.702309] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c3abd0-6526-4925-8f32-704627ec8be1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.714371] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7c37c6-25ec-4ca1-b9b0-248c8fa80b40 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.717302] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263941, 'name': Rename_Task, 'duration_secs': 0.138852} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.717568] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1136.718094] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d2670ff8-434a-4883-81cc-1a7fcccbfb08 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.745287] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c757a65-5af1-446a-b167-94001125f3fb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.747841] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1136.747841] env[62952]: value = "task-1263942" [ 1136.747841] env[62952]: _type = "Task" [ 1136.747841] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.753811] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8708c29f-7b47-4db5-8c44-6e7b32c0252d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.762216] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263942, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.772215] env[62952]: DEBUG nova.compute.provider_tree [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1136.978793] env[62952]: DEBUG nova.objects.base [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Object Instance<11fcbf44-5b9b-4762-8be8-f4e1d0cfa278> lazy-loaded attributes: info_cache,migration_context {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1136.979787] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-253bffb4-fb59-4a61-bda7-bf331e8f268a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.999358] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d64673a-70b1-4ada-9ca5-9c7d1d9c316f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.004589] env[62952]: DEBUG oslo_vmware.api [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1137.004589] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52341a71-9647-7c3d-4c49-93e0b40683d8" [ 1137.004589] env[62952]: _type = "Task" [ 1137.004589] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.012167] env[62952]: DEBUG oslo_vmware.api [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52341a71-9647-7c3d-4c49-93e0b40683d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.257988] env[62952]: DEBUG oslo_vmware.api [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263942, 'name': PowerOnVM_Task, 'duration_secs': 0.416306} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.258383] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1137.275700] env[62952]: DEBUG nova.scheduler.client.report [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1137.353869] env[62952]: DEBUG nova.compute.manager [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1137.354821] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15120963-0332-4957-a5f3-fc37e5a408b9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.514705] env[62952]: DEBUG oslo_vmware.api [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52341a71-9647-7c3d-4c49-93e0b40683d8, 'name': SearchDatastore_Task, 'duration_secs': 0.006476} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.514982] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.783372] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.689s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.785818] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.271s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.812104] env[62952]: INFO nova.scheduler.client.report [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Deleted allocations for instance 1941365e-d562-4d99-89eb-226a07e52071 [ 1137.870804] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142dc24f-4e6d-4476-85c6-96879a82a372 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "42793459-89e6-4941-9b20-cbe52a241c6d" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 22.061s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.305932] env[62952]: DEBUG nova.scheduler.client.report [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Refreshing inventories for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1138.319230] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e529c728-6eea-4713-9da2-fad4670f77a7 tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "1941365e-d562-4d99-89eb-226a07e52071" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.802s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.321294] env[62952]: DEBUG nova.scheduler.client.report [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Updating ProviderTree inventory for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1138.321514] env[62952]: DEBUG nova.compute.provider_tree [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1138.333047] env[62952]: DEBUG nova.scheduler.client.report [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Refreshing aggregate associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1138.350381] env[62952]: DEBUG nova.scheduler.client.report [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Refreshing trait associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1138.433934] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa953d9-27b7-4850-9132-49f1beab28af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.441596] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-debb6dcd-5fe0-489d-a004-24413621f9ea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.472919] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-995c8ba5-757b-46ea-9ecd-c6367cbfc0cb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.479993] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe30056-5731-4e86-b1fc-21764f553442 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.492735] env[62952]: DEBUG nova.compute.provider_tree [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1138.997052] env[62952]: DEBUG nova.scheduler.client.report [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1140.006698] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.221s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.565839] env[62952]: INFO nova.scheduler.client.report [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleted allocation for migration 1cead09a-f727-4baf-84e9-9e025e5ff184 [ 1141.071163] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.882s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.213861] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.214165] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.214410] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.214607] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.214783] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.217297] env[62952]: INFO nova.compute.manager [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Terminating instance [ 1142.219014] env[62952]: DEBUG nova.compute.manager [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1142.219232] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1142.220126] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f9f6d5-030b-4734-95bb-96074af50451 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.227362] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1142.227606] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-edac4055-e481-436f-855b-1626e45870bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.233257] env[62952]: DEBUG oslo_vmware.api [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1142.233257] env[62952]: value = "task-1263944" [ 1142.233257] env[62952]: _type = "Task" [ 1142.233257] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.240748] env[62952]: DEBUG oslo_vmware.api [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263944, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.743657] env[62952]: DEBUG oslo_vmware.api [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263944, 'name': PowerOffVM_Task, 'duration_secs': 0.18821} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.743954] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1142.744154] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1142.744441] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e69d3bed-0c34-4322-ba59-e8393f2299a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.802813] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1142.803272] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1142.803272] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleting the datastore file [datastore1] 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1142.803517] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4d3515ca-c8c2-42bc-a2eb-a51d4c0e67f1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.809555] env[62952]: DEBUG oslo_vmware.api [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for the task: (returnval){ [ 1142.809555] env[62952]: value = "task-1263946" [ 1142.809555] env[62952]: _type = "Task" [ 1142.809555] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.817656] env[62952]: DEBUG oslo_vmware.api [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263946, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.319740] env[62952]: DEBUG oslo_vmware.api [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Task: {'id': task-1263946, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13786} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.320136] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1143.320214] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1143.320385] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1143.320582] env[62952]: INFO nova.compute.manager [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1143.320888] env[62952]: DEBUG oslo.service.loopingcall [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1143.321112] env[62952]: DEBUG nova.compute.manager [-] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1143.321211] env[62952]: DEBUG nova.network.neutron [-] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1143.574543] env[62952]: DEBUG nova.compute.manager [req-8228513e-1f4d-4f5f-bf5f-5f6e4a5045a8 req-b3ee2413-c134-47e9-aa43-2b811b3f0c14 service nova] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Received event network-vif-deleted-9d5ab9ea-3d2a-40d3-996d-96b27d38cff5 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1143.574769] env[62952]: INFO nova.compute.manager [req-8228513e-1f4d-4f5f-bf5f-5f6e4a5045a8 req-b3ee2413-c134-47e9-aa43-2b811b3f0c14 service nova] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Neutron deleted interface 9d5ab9ea-3d2a-40d3-996d-96b27d38cff5; detaching it from the instance and deleting it from the info cache [ 1143.574934] env[62952]: DEBUG nova.network.neutron [req-8228513e-1f4d-4f5f-bf5f-5f6e4a5045a8 req-b3ee2413-c134-47e9-aa43-2b811b3f0c14 service nova] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.785551] env[62952]: DEBUG oslo_concurrency.lockutils [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "77fe79de-8da8-486e-af63-91f9d8196ca9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.785824] env[62952]: DEBUG oslo_concurrency.lockutils [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "77fe79de-8da8-486e-af63-91f9d8196ca9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.786054] env[62952]: DEBUG oslo_concurrency.lockutils [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "77fe79de-8da8-486e-af63-91f9d8196ca9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.786249] env[62952]: DEBUG oslo_concurrency.lockutils [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "77fe79de-8da8-486e-af63-91f9d8196ca9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.786466] env[62952]: DEBUG oslo_concurrency.lockutils [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "77fe79de-8da8-486e-af63-91f9d8196ca9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.788777] env[62952]: INFO nova.compute.manager [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Terminating instance [ 1143.790385] env[62952]: DEBUG nova.compute.manager [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1143.790595] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1143.791466] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015248e3-6081-497d-9a62-a7441ac88e24 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.799263] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1143.799450] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-98b3e48b-6ef6-42ed-a151-a897df2d1a88 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.805882] env[62952]: DEBUG oslo_vmware.api [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1143.805882] env[62952]: value = "task-1263947" [ 1143.805882] env[62952]: _type = "Task" [ 1143.805882] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.812912] env[62952]: DEBUG oslo_vmware.api [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263947, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.054740] env[62952]: DEBUG nova.network.neutron [-] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.077655] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cfc9c289-da25-4ea1-9d6a-55311e1df473 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.086320] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ac4842-119e-4399-96cf-d297bdb39b1a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.112144] env[62952]: DEBUG nova.compute.manager [req-8228513e-1f4d-4f5f-bf5f-5f6e4a5045a8 req-b3ee2413-c134-47e9-aa43-2b811b3f0c14 service nova] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Detach interface failed, port_id=9d5ab9ea-3d2a-40d3-996d-96b27d38cff5, reason: Instance 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1144.315673] env[62952]: DEBUG oslo_vmware.api [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263947, 'name': PowerOffVM_Task, 'duration_secs': 0.234414} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.315966] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1144.316160] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1144.316451] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ebe87086-303e-4236-8684-5327f3ae8787 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.385019] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1144.385455] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1144.385698] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Deleting the datastore file [datastore1] 77fe79de-8da8-486e-af63-91f9d8196ca9 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1144.386058] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-802b0675-3f10-4373-82f0-89953551517d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.392541] env[62952]: DEBUG oslo_vmware.api [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1144.392541] env[62952]: value = "task-1263949" [ 1144.392541] env[62952]: _type = "Task" [ 1144.392541] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.400719] env[62952]: DEBUG oslo_vmware.api [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263949, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.557706] env[62952]: INFO nova.compute.manager [-] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Took 1.24 seconds to deallocate network for instance. [ 1144.903012] env[62952]: DEBUG oslo_vmware.api [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263949, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159625} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.903303] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1144.903494] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1144.903681] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1144.903855] env[62952]: INFO nova.compute.manager [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1144.904109] env[62952]: DEBUG oslo.service.loopingcall [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1144.904310] env[62952]: DEBUG nova.compute.manager [-] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1144.904406] env[62952]: DEBUG nova.network.neutron [-] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1145.065421] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.065713] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.065992] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.100973] env[62952]: INFO nova.scheduler.client.report [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Deleted allocations for instance 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278 [ 1145.272066] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1145.272247] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Cleaning up deleted instances with incomplete migration {{(pid=62952) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1145.603428] env[62952]: DEBUG nova.compute.manager [req-9f36defe-e868-48dd-a7c8-6fb9c9b799e8 req-b9a5c0b4-2a3b-4176-b6f1-1e610dbeece8 service nova] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Received event network-vif-deleted-bd2fc6b2-78ee-4677-899a-db20d7a7ce8f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1145.603659] env[62952]: INFO nova.compute.manager [req-9f36defe-e868-48dd-a7c8-6fb9c9b799e8 req-b9a5c0b4-2a3b-4176-b6f1-1e610dbeece8 service nova] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Neutron deleted interface bd2fc6b2-78ee-4677-899a-db20d7a7ce8f; detaching it from the instance and deleting it from the info cache [ 1145.603809] env[62952]: DEBUG nova.network.neutron [req-9f36defe-e868-48dd-a7c8-6fb9c9b799e8 req-b9a5c0b4-2a3b-4176-b6f1-1e610dbeece8 service nova] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.608581] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91691b04-2f7d-41f2-b30d-51f8a984a5c2 tempest-DeleteServersTestJSON-1019128451 tempest-DeleteServersTestJSON-1019128451-project-member] Lock "11fcbf44-5b9b-4762-8be8-f4e1d0cfa278" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.394s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.663524] env[62952]: DEBUG nova.network.neutron [-] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.106770] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dfb99f86-0199-40b6-9e0d-116d47d2d287 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.115762] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a4baa4-6d6e-483b-b2f5-9fff1dbcb082 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.142570] env[62952]: DEBUG nova.compute.manager [req-9f36defe-e868-48dd-a7c8-6fb9c9b799e8 req-b9a5c0b4-2a3b-4176-b6f1-1e610dbeece8 service nova] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Detach interface failed, port_id=bd2fc6b2-78ee-4677-899a-db20d7a7ce8f, reason: Instance 77fe79de-8da8-486e-af63-91f9d8196ca9 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1146.165323] env[62952]: INFO nova.compute.manager [-] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Took 1.26 seconds to deallocate network for instance. [ 1146.672344] env[62952]: DEBUG oslo_concurrency.lockutils [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.672646] env[62952]: DEBUG oslo_concurrency.lockutils [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.672940] env[62952]: DEBUG nova.objects.instance [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lazy-loading 'resources' on Instance uuid 77fe79de-8da8-486e-af63-91f9d8196ca9 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1146.774549] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1146.774849] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1146.790307] env[62952]: DEBUG nova.compute.manager [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Stashing vm_state: active {{(pid=62952) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1147.261082] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d3f031b-19e6-4f29-be6d-f4103b6774c4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.269140] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dd38540-4ea7-4d85-bb85-475cda3a23bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.299016] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.302755] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a98515d4-3518-412b-ac34-7140a12ccb3c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.310500] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66cd0e50-2782-4fa8-af56-c1040e468526 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.314848] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.324980] env[62952]: DEBUG nova.compute.provider_tree [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1147.828855] env[62952]: DEBUG nova.scheduler.client.report [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1148.333891] env[62952]: DEBUG oslo_concurrency.lockutils [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.661s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.336203] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.037s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.336430] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.336590] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1148.336922] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.022s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.339073] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b156b82-51f4-40fe-8970-0d64dff3f6cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.347702] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc97e6e-0ea8-4b69-acb9-1481aec91732 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.352590] env[62952]: INFO nova.scheduler.client.report [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Deleted allocations for instance 77fe79de-8da8-486e-af63-91f9d8196ca9 [ 1148.366687] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14fd562d-db42-475d-bcd7-105f3bc119da {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.373180] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612d47bd-b944-4331-89a8-83278a7be34f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.402145] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180891MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1148.402345] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.845024] env[62952]: INFO nova.compute.claims [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1148.859281] env[62952]: DEBUG oslo_concurrency.lockutils [None req-983c884c-a3e7-4b8d-bb58-4020d7d56dbc tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "77fe79de-8da8-486e-af63-91f9d8196ca9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.073s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.320060] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "dbca4369-1ed3-493a-9847-9fa1e4293475" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.320060] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "dbca4369-1ed3-493a-9847-9fa1e4293475" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.320060] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "dbca4369-1ed3-493a-9847-9fa1e4293475-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.320310] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "dbca4369-1ed3-493a-9847-9fa1e4293475-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.320383] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "dbca4369-1ed3-493a-9847-9fa1e4293475-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.322607] env[62952]: INFO nova.compute.manager [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Terminating instance [ 1149.324847] env[62952]: DEBUG nova.compute.manager [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1149.325079] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1149.325928] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77cab656-2eed-46cf-b891-a337ead25ba5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.333540] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1149.333831] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-55280a89-460b-4097-9c28-3af7a33618e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.339437] env[62952]: DEBUG oslo_vmware.api [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1149.339437] env[62952]: value = "task-1263951" [ 1149.339437] env[62952]: _type = "Task" [ 1149.339437] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.347315] env[62952]: DEBUG oslo_vmware.api [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263951, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.349957] env[62952]: INFO nova.compute.resource_tracker [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating resource usage from migration 5ec1a69b-8197-4f3d-b519-246ded550e47 [ 1149.423024] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efce2981-73a7-47b5-b988-bf16bec0f057 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.429861] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31538c9a-9c8d-4917-b8c7-8c6c5cd018d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.460667] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1202c4d-3904-4bdb-9c79-3852c82718a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.468344] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da16c486-27eb-4757-9a4a-47cec7a55ab5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.481615] env[62952]: DEBUG nova.compute.provider_tree [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1149.851759] env[62952]: DEBUG oslo_vmware.api [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263951, 'name': PowerOffVM_Task, 'duration_secs': 0.21374} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.852147] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1149.852216] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1149.852461] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9bdfc159-e9c8-4578-8899-532cb6f84130 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.916072] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1149.916348] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1149.916549] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Deleting the datastore file [datastore1] dbca4369-1ed3-493a-9847-9fa1e4293475 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1149.916839] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e08add83-36af-462c-b27f-5dc3620bf1bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.923795] env[62952]: DEBUG oslo_vmware.api [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for the task: (returnval){ [ 1149.923795] env[62952]: value = "task-1263953" [ 1149.923795] env[62952]: _type = "Task" [ 1149.923795] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.931465] env[62952]: DEBUG oslo_vmware.api [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263953, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.985026] env[62952]: DEBUG nova.scheduler.client.report [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1150.433999] env[62952]: DEBUG oslo_vmware.api [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Task: {'id': task-1263953, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137048} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.434320] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1150.434520] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1150.434702] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1150.434881] env[62952]: INFO nova.compute.manager [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1150.435144] env[62952]: DEBUG oslo.service.loopingcall [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1150.435346] env[62952]: DEBUG nova.compute.manager [-] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1150.435445] env[62952]: DEBUG nova.network.neutron [-] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1150.490672] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.153s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.490672] env[62952]: INFO nova.compute.manager [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Migrating [ 1150.497266] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.095s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.742166] env[62952]: DEBUG nova.compute.manager [req-924a511a-4614-46ad-9c56-c4f10624e117 req-b8b62e51-7391-420f-9b9b-76562532a334 service nova] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Received event network-vif-deleted-d93fe914-641a-4750-a765-181c7012b252 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1150.743090] env[62952]: INFO nova.compute.manager [req-924a511a-4614-46ad-9c56-c4f10624e117 req-b8b62e51-7391-420f-9b9b-76562532a334 service nova] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Neutron deleted interface d93fe914-641a-4750-a765-181c7012b252; detaching it from the instance and deleting it from the info cache [ 1150.743090] env[62952]: DEBUG nova.network.neutron [req-924a511a-4614-46ad-9c56-c4f10624e117 req-b8b62e51-7391-420f-9b9b-76562532a334 service nova] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1151.009388] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1151.009668] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.009757] env[62952]: DEBUG nova.network.neutron [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1151.218925] env[62952]: DEBUG nova.network.neutron [-] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1151.245022] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ff573dc-85e7-47a5-80a4-896b946fb907 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.256060] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fd9a979-8f64-4498-b375-019f0d065cb6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.282398] env[62952]: DEBUG nova.compute.manager [req-924a511a-4614-46ad-9c56-c4f10624e117 req-b8b62e51-7391-420f-9b9b-76562532a334 service nova] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Detach interface failed, port_id=d93fe914-641a-4750-a765-181c7012b252, reason: Instance dbca4369-1ed3-493a-9847-9fa1e4293475 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1151.514724] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Applying migration context for instance 43d79d26-4350-4329-bffc-cedfada40ee0 as it has an incoming, in-progress migration 5ec1a69b-8197-4f3d-b519-246ded550e47. Migration status is pre-migrating {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1151.515561] env[62952]: INFO nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating resource usage from migration 5ec1a69b-8197-4f3d-b519-246ded550e47 [ 1151.538686] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance dbca4369-1ed3-493a-9847-9fa1e4293475 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1151.538686] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance f92ef782-fd64-47f0-8145-9261a1c3005c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1151.538874] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 42793459-89e6-4941-9b20-cbe52a241c6d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1151.538874] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Migration 5ec1a69b-8197-4f3d-b519-246ded550e47 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1151.538955] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 43d79d26-4350-4329-bffc-cedfada40ee0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1151.539170] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1151.539307] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1536MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1151.605119] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6d10bd-8c6a-4be0-bfce-53cbd91fb717 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.614344] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e68a190c-1bbb-469e-9d1d-a422228e7912 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.645548] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71bc5e96-bd24-4d02-92b6-06eb590af264 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.652565] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da8b531e-f812-4cad-a0ac-f4400044c395 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.665502] env[62952]: DEBUG nova.compute.provider_tree [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1151.722256] env[62952]: INFO nova.compute.manager [-] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Took 1.29 seconds to deallocate network for instance. [ 1151.744629] env[62952]: DEBUG nova.network.neutron [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating instance_info_cache with network_info: [{"id": "e6fdb64f-1d51-4b10-b297-9f760ae81c95", "address": "fa:16:3e:95:ff:4c", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6fdb64f-1d", "ovs_interfaceid": "e6fdb64f-1d51-4b10-b297-9f760ae81c95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1152.169011] env[62952]: DEBUG nova.scheduler.client.report [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1152.228954] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.246728] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1152.674068] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1152.674312] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.177s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.674606] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.446s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.674843] env[62952]: DEBUG nova.objects.instance [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lazy-loading 'resources' on Instance uuid dbca4369-1ed3-493a-9847-9fa1e4293475 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1152.676025] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1152.676128] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Cleaning up deleted instances {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1153.192543] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] There are 52 instances to clean {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1153.192809] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 148e52e8-07ef-4b3d-a43b-eefc6820b248] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1153.382789] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-074db000-6ba2-497e-a9e8-3a8bbceee175 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.390422] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af92f5c-7fa3-4a25-a3d3-193c531d7403 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.419990] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b2b0b8-55e5-42d1-9ed4-9a2e8a32fa96 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.426669] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3771d6-be07-4f8d-8b9e-9316cbcfb614 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.439248] env[62952]: DEBUG nova.compute.provider_tree [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1153.695585] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 11fcbf44-5b9b-4762-8be8-f4e1d0cfa278] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1153.761746] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f2eecbc-a96c-47aa-a7e4-7b9132e53573 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.781541] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating instance '43d79d26-4350-4329-bffc-cedfada40ee0' progress to 0 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1153.943074] env[62952]: DEBUG nova.scheduler.client.report [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1154.199107] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 126f57b5-be03-4164-b92f-3c2e96dd9171] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1154.287548] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1154.287677] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7491ff24-a5fb-4d66-baba-6645695556f8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.294913] env[62952]: DEBUG oslo_vmware.api [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1154.294913] env[62952]: value = "task-1263954" [ 1154.294913] env[62952]: _type = "Task" [ 1154.294913] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.302644] env[62952]: DEBUG oslo_vmware.api [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263954, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.448559] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.774s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.470336] env[62952]: INFO nova.scheduler.client.report [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Deleted allocations for instance dbca4369-1ed3-493a-9847-9fa1e4293475 [ 1154.702287] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 8cda6eea-d488-4d73-a7ab-7a905037929d] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1154.805191] env[62952]: DEBUG oslo_vmware.api [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263954, 'name': PowerOffVM_Task, 'duration_secs': 0.191842} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.805501] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1154.805698] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating instance '43d79d26-4350-4329-bffc-cedfada40ee0' progress to 17 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1154.978129] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0880f493-3e79-4d7e-9752-6a3753143dea tempest-ServerRescueNegativeTestJSON-1392265834 tempest-ServerRescueNegativeTestJSON-1392265834-project-member] Lock "dbca4369-1ed3-493a-9847-9fa1e4293475" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.658s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.205518] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: f28d0043-236e-4df4-8478-e4b1befb6a7d] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1155.312579] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1155.312768] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1155.312928] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1155.313127] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1155.313280] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1155.313434] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1155.313635] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1155.313798] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1155.313965] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1155.314159] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1155.314344] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1155.319393] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d430d3e1-bf88-429f-a4d3-6971ab549a71 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.334620] env[62952]: DEBUG oslo_vmware.api [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1155.334620] env[62952]: value = "task-1263955" [ 1155.334620] env[62952]: _type = "Task" [ 1155.334620] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.341947] env[62952]: DEBUG oslo_vmware.api [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263955, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.708512] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 1a02fba4-aad0-4bd4-8183-a7f08d887a5a] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1155.844504] env[62952]: DEBUG oslo_vmware.api [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263955, 'name': ReconfigVM_Task, 'duration_secs': 0.154292} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.845158] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating instance '43d79d26-4350-4329-bffc-cedfada40ee0' progress to 33 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1156.211228] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: af261b32-457a-44a1-bab4-b62654b0be1f] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1156.350999] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1156.350999] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1156.351268] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1156.351391] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1156.351578] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1156.351736] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1156.351996] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1156.352090] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1156.352352] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1156.352430] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1156.352683] env[62952]: DEBUG nova.virt.hardware [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1156.358386] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1156.359056] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fed6f09-a25b-4407-9421-d73bba435d17 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.377473] env[62952]: DEBUG oslo_vmware.api [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1156.377473] env[62952]: value = "task-1263956" [ 1156.377473] env[62952]: _type = "Task" [ 1156.377473] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.385870] env[62952]: DEBUG oslo_vmware.api [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263956, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.715952] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 1a38bdcd-5537-46f5-94f6-a8e358e13121] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1156.888526] env[62952]: DEBUG oslo_vmware.api [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263956, 'name': ReconfigVM_Task, 'duration_secs': 0.163633} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.889194] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1156.889726] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de17e793-d006-4460-b8ce-6dbb7c5fe4e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.912298] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 43d79d26-4350-4329-bffc-cedfada40ee0/43d79d26-4350-4329-bffc-cedfada40ee0.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1156.912576] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-023a7e2a-6a36-4da4-9659-4be396a33cab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.930302] env[62952]: DEBUG oslo_vmware.api [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1156.930302] env[62952]: value = "task-1263957" [ 1156.930302] env[62952]: _type = "Task" [ 1156.930302] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.938181] env[62952]: DEBUG oslo_vmware.api [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263957, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.218715] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: b1c9a7c3-a405-4519-8d2e-a3d9bb50a76b] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1157.439357] env[62952]: DEBUG oslo_vmware.api [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263957, 'name': ReconfigVM_Task, 'duration_secs': 0.246212} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.439732] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 43d79d26-4350-4329-bffc-cedfada40ee0/43d79d26-4350-4329-bffc-cedfada40ee0.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1157.440089] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating instance '43d79d26-4350-4329-bffc-cedfada40ee0' progress to 50 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1157.721698] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: e9e690b8-2fae-401d-aaad-0acef6e88ce3] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1157.955242] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f58995b-88ed-47d6-8b6c-27757e41dfa6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.976481] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a1b41a-a7ab-4e48-8027-9d483997b0bc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.995431] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating instance '43d79d26-4350-4329-bffc-cedfada40ee0' progress to 67 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1158.225834] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 05223053-90c3-433c-99b9-e252ef9a1935] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1158.536869] env[62952]: DEBUG nova.network.neutron [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Port e6fdb64f-1d51-4b10-b297-9f760ae81c95 binding to destination host cpu-1 is already ACTIVE {{(pid=62952) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1158.729835] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 4d83faf9-43db-4fc4-978b-55ca9a6ed4c9] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1159.233371] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: d98a88aa-be5f-4254-aaba-c6fa03c9269a] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1159.351539] env[62952]: DEBUG nova.compute.manager [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1159.352451] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b566829b-8fab-441f-a5a8-96157341d415 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.557854] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "43d79d26-4350-4329-bffc-cedfada40ee0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.557854] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "43d79d26-4350-4329-bffc-cedfada40ee0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.558096] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "43d79d26-4350-4329-bffc-cedfada40ee0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.736341] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 1941365e-d562-4d99-89eb-226a07e52071] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1159.862490] env[62952]: INFO nova.compute.manager [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] instance snapshotting [ 1159.863159] env[62952]: DEBUG nova.objects.instance [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lazy-loading 'flavor' on Instance uuid f92ef782-fd64-47f0-8145-9261a1c3005c {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1160.239951] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 511b995d-c531-41f5-9d1b-89ec8c1bfc18] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1160.368658] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c3c1e0-d744-4ade-b2fa-259fd485f193 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.387477] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a18b0ce0-d9f1-48fe-a235-6b1109d959f1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.592070] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.592298] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.592479] env[62952]: DEBUG nova.network.neutron [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1160.743683] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 120c6804-5aab-4406-9c0c-46eaa835b59f] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1160.897408] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1160.897735] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d1cae3a4-0c92-45f8-8718-0098ac797fca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.904931] env[62952]: DEBUG oslo_vmware.api [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1160.904931] env[62952]: value = "task-1263958" [ 1160.904931] env[62952]: _type = "Task" [ 1160.904931] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.912619] env[62952]: DEBUG oslo_vmware.api [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263958, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.246792] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: afaecd3c-da24-4111-8262-d92a57a42a55] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1161.328150] env[62952]: DEBUG nova.network.neutron [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating instance_info_cache with network_info: [{"id": "e6fdb64f-1d51-4b10-b297-9f760ae81c95", "address": "fa:16:3e:95:ff:4c", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6fdb64f-1d", "ovs_interfaceid": "e6fdb64f-1d51-4b10-b297-9f760ae81c95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1161.414861] env[62952]: DEBUG oslo_vmware.api [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263958, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.750588] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: e44914cd-d8ae-4f43-9207-32cacf6d6553] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1161.830355] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.915507] env[62952]: DEBUG oslo_vmware.api [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263958, 'name': CreateSnapshot_Task, 'duration_secs': 0.532685} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.915781] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1161.916563] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25f3cfc6-43c8-435c-aebc-983aafef9ce1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.254243] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 6f0b04f5-811f-4c53-808b-6d9d22100a86] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1162.355018] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e86bcc-3f12-4b3b-84ac-cd55e205a274 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.373784] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4991db69-6fd2-4d51-bc4d-0413fe696a49 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.380107] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating instance '43d79d26-4350-4329-bffc-cedfada40ee0' progress to 83 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1162.435182] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1162.435678] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-497d18c1-1bca-41c5-9d25-b5d04057c8c8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.444561] env[62952]: DEBUG oslo_vmware.api [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1162.444561] env[62952]: value = "task-1263959" [ 1162.444561] env[62952]: _type = "Task" [ 1162.444561] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.452472] env[62952]: DEBUG oslo_vmware.api [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263959, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.757724] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 7ba1587d-edaf-4dce-a224-6b2ac22cfecd] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1162.885907] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1162.886277] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-814d7244-ad60-4e56-abbe-3f482f8a3c56 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.893415] env[62952]: DEBUG oslo_vmware.api [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1162.893415] env[62952]: value = "task-1263960" [ 1162.893415] env[62952]: _type = "Task" [ 1162.893415] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.901017] env[62952]: DEBUG oslo_vmware.api [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263960, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.954826] env[62952]: DEBUG oslo_vmware.api [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263959, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.261317] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 75ee70f6-e12c-4c21-ac57-1b07b5d8a4a8] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1163.403021] env[62952]: DEBUG oslo_vmware.api [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263960, 'name': PowerOnVM_Task, 'duration_secs': 0.438631} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.403318] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1163.403510] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c7d7edf5-0979-4202-8260-4ca9a8adf40f tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating instance '43d79d26-4350-4329-bffc-cedfada40ee0' progress to 100 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1163.460256] env[62952]: DEBUG oslo_vmware.api [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263959, 'name': CloneVM_Task, 'duration_secs': 0.93288} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.460599] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Created linked-clone VM from snapshot [ 1163.461650] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b04e9c-e721-4983-869e-bc6904e5b9b6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.471070] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Uploading image efbe8091-a483-4979-a9f8-af2c3c2a058c {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1163.494535] env[62952]: DEBUG oslo_vmware.rw_handles [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1163.494535] env[62952]: value = "vm-271985" [ 1163.494535] env[62952]: _type = "VirtualMachine" [ 1163.494535] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1163.495155] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1f49c004-c00f-4701-aec9-d249791f99a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.501713] env[62952]: DEBUG oslo_vmware.rw_handles [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lease: (returnval){ [ 1163.501713] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d719ed-1363-2a3b-05be-e2fa47d96578" [ 1163.501713] env[62952]: _type = "HttpNfcLease" [ 1163.501713] env[62952]: } obtained for exporting VM: (result){ [ 1163.501713] env[62952]: value = "vm-271985" [ 1163.501713] env[62952]: _type = "VirtualMachine" [ 1163.501713] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1163.502206] env[62952]: DEBUG oslo_vmware.api [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the lease: (returnval){ [ 1163.502206] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d719ed-1363-2a3b-05be-e2fa47d96578" [ 1163.502206] env[62952]: _type = "HttpNfcLease" [ 1163.502206] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1163.509323] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1163.509323] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d719ed-1363-2a3b-05be-e2fa47d96578" [ 1163.509323] env[62952]: _type = "HttpNfcLease" [ 1163.509323] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1163.766066] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: c7d08234-fcef-4cdb-939f-05ea9e8b08ea] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1164.010030] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1164.010030] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d719ed-1363-2a3b-05be-e2fa47d96578" [ 1164.010030] env[62952]: _type = "HttpNfcLease" [ 1164.010030] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1164.010280] env[62952]: DEBUG oslo_vmware.rw_handles [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1164.010280] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d719ed-1363-2a3b-05be-e2fa47d96578" [ 1164.010280] env[62952]: _type = "HttpNfcLease" [ 1164.010280] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1164.010973] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-045de246-dfaf-447a-8cb1-b8b8a9516bc0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.018119] env[62952]: DEBUG oslo_vmware.rw_handles [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52286f39-46dc-f5a8-94ea-ac42eb4337f2/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1164.018269] env[62952]: DEBUG oslo_vmware.rw_handles [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52286f39-46dc-f5a8-94ea-ac42eb4337f2/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1164.103161] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b7dc0cee-f030-41f7-b83f-1670324b6b6e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.268401] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: fba2e955-8b5d-461c-997b-ef8f3ec0d906] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1164.772431] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 77fe79de-8da8-486e-af63-91f9d8196ca9] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1165.275736] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: b0e3ed5f-5e80-4c40-ba6d-29b7a5466e6e] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1165.455228] env[62952]: DEBUG nova.network.neutron [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Port e6fdb64f-1d51-4b10-b297-9f760ae81c95 binding to destination host cpu-1 is already ACTIVE {{(pid=62952) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1165.455589] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1165.455921] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.456178] env[62952]: DEBUG nova.network.neutron [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1165.779814] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 20ea323b-8961-4981-8045-f9cb85815d99] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1166.176706] env[62952]: DEBUG nova.network.neutron [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating instance_info_cache with network_info: [{"id": "e6fdb64f-1d51-4b10-b297-9f760ae81c95", "address": "fa:16:3e:95:ff:4c", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6fdb64f-1d", "ovs_interfaceid": "e6fdb64f-1d51-4b10-b297-9f760ae81c95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.282649] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: e17b3d83-7cea-4a65-9457-7e57a6e7641a] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1166.680228] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1166.785693] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: dafa4926-3ab5-47f9-8ff2-4049e05c3113] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1167.185147] env[62952]: DEBUG nova.compute.manager [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62952) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1167.185549] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1167.185926] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1167.290052] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 45a6cbf4-dea1-4782-bf4d-a40ff6fa2a20] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1167.689409] env[62952]: DEBUG nova.objects.instance [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lazy-loading 'migration_context' on Instance uuid 43d79d26-4350-4329-bffc-cedfada40ee0 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1167.796360] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 7a73cb49-3e93-4fcc-b613-3808a8d5c64a] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1168.293344] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25fc4fe0-ccbe-4215-816d-98b8bf21ae7b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.300050] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 051f9279-533a-4545-b9e2-272bf40956f6] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1168.302661] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b652ea-589c-4e1a-934c-771523aa6ebb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.334753] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05cfe12b-45f0-4cee-ad19-07d2fda376c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.342592] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f91d5ca-0639-488d-bd0b-51a22773ff85 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.356421] env[62952]: DEBUG nova.compute.provider_tree [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1168.806778] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: e5e7b2b5-34cf-4d37-87ff-1b57264a3516] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1168.859984] env[62952]: DEBUG nova.scheduler.client.report [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1169.310028] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 47b57348-40d1-4735-9f5a-a7288f8f12ee] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1169.813649] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 86ec7b3e-34d6-45e8-87b2-c8550f035ed0] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1169.870945] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.685s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.317341] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 48fcd1c2-eccd-4d8c-a0f1-ec7ecbd280f5] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1170.820733] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 14f8dbbe-2c4a-4948-bf15-106d5b1b0677] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1171.324255] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: ba022bea-e357-4541-a573-d1da2eb49b46] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1171.415090] env[62952]: INFO nova.compute.manager [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Swapping old allocation on dict_keys(['615127b5-dc82-4491-9c8f-4e6d0caa0690']) held by migration 5ec1a69b-8197-4f3d-b519-246ded550e47 for instance [ 1171.438665] env[62952]: DEBUG nova.scheduler.client.report [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Overwriting current allocation {'allocations': {'615127b5-dc82-4491-9c8f-4e6d0caa0690': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 136}}, 'project_id': '5080fbeac0744478843ab28efea8fc18', 'user_id': 'ffc122d64b19432aae8d9627d2928c06', 'consumer_generation': 1} on consumer 43d79d26-4350-4329-bffc-cedfada40ee0 {{(pid=62952) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1171.516906] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1171.517132] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.517330] env[62952]: DEBUG nova.network.neutron [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1171.827660] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 49947bde-069b-459c-ae19-ae5f090f535b] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1172.263462] env[62952]: DEBUG nova.network.neutron [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating instance_info_cache with network_info: [{"id": "e6fdb64f-1d51-4b10-b297-9f760ae81c95", "address": "fa:16:3e:95:ff:4c", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6fdb64f-1d", "ovs_interfaceid": "e6fdb64f-1d51-4b10-b297-9f760ae81c95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.327060] env[62952]: DEBUG oslo_vmware.rw_handles [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52286f39-46dc-f5a8-94ea-ac42eb4337f2/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1172.328103] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672f1de1-6b0a-4d5f-93c7-8bba896289d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.330789] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 3416eb38-b961-4d48-8b37-8b7b92375dc9] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1172.335992] env[62952]: DEBUG oslo_vmware.rw_handles [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52286f39-46dc-f5a8-94ea-ac42eb4337f2/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1172.336209] env[62952]: ERROR oslo_vmware.rw_handles [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52286f39-46dc-f5a8-94ea-ac42eb4337f2/disk-0.vmdk due to incomplete transfer. [ 1172.336438] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c58e14ac-64f4-42b5-9b63-245341f0d6ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.343559] env[62952]: DEBUG oslo_vmware.rw_handles [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52286f39-46dc-f5a8-94ea-ac42eb4337f2/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1172.343751] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Uploaded image efbe8091-a483-4979-a9f8-af2c3c2a058c to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1172.346139] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1172.346365] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-311759e7-7dbd-4fca-be0b-906408b33de4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.351675] env[62952]: DEBUG oslo_vmware.api [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1172.351675] env[62952]: value = "task-1263962" [ 1172.351675] env[62952]: _type = "Task" [ 1172.351675] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.360061] env[62952]: DEBUG oslo_vmware.api [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263962, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.766776] env[62952]: DEBUG oslo_concurrency.lockutils [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "refresh_cache-43d79d26-4350-4329-bffc-cedfada40ee0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1172.767374] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1172.767730] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cda0159f-f6a5-4043-a9ff-70289d365a98 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.775220] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1172.775220] env[62952]: value = "task-1263963" [ 1172.775220] env[62952]: _type = "Task" [ 1172.775220] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.782907] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263963, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.834598] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: dd700ee9-0957-4210-be50-e4c8ac3ca456] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1172.861630] env[62952]: DEBUG oslo_vmware.api [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263962, 'name': Destroy_Task, 'duration_secs': 0.292679} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.861914] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Destroyed the VM [ 1172.862176] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1172.862435] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-490b78a6-ed89-418f-ba25-fda53ab1ef09 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.869220] env[62952]: DEBUG oslo_vmware.api [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1172.869220] env[62952]: value = "task-1263964" [ 1172.869220] env[62952]: _type = "Task" [ 1172.869220] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.877286] env[62952]: DEBUG oslo_vmware.api [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263964, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.286069] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263963, 'name': PowerOffVM_Task, 'duration_secs': 0.219831} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.286200] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1173.286839] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1173.287102] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1173.287284] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1173.287477] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1173.287627] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1173.287775] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1173.287982] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1173.288161] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1173.288336] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1173.288503] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1173.288681] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1173.293841] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62bfeb70-91fd-41a5-847e-5e7d78f26fe8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.309115] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1173.309115] env[62952]: value = "task-1263965" [ 1173.309115] env[62952]: _type = "Task" [ 1173.309115] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.317186] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263965, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.337793] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 6e289c20-8dd2-4680-b816-9bbe82fe7aec] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1173.379587] env[62952]: DEBUG oslo_vmware.api [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263964, 'name': RemoveSnapshot_Task, 'duration_secs': 0.469636} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.379850] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1173.380084] env[62952]: INFO nova.compute.manager [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Took 13.01 seconds to snapshot the instance on the hypervisor. [ 1173.520314] env[62952]: DEBUG oslo_concurrency.lockutils [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "42793459-89e6-4941-9b20-cbe52a241c6d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.520594] env[62952]: DEBUG oslo_concurrency.lockutils [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "42793459-89e6-4941-9b20-cbe52a241c6d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.520809] env[62952]: DEBUG oslo_concurrency.lockutils [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "42793459-89e6-4941-9b20-cbe52a241c6d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.521012] env[62952]: DEBUG oslo_concurrency.lockutils [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "42793459-89e6-4941-9b20-cbe52a241c6d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.521221] env[62952]: DEBUG oslo_concurrency.lockutils [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "42793459-89e6-4941-9b20-cbe52a241c6d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1173.523323] env[62952]: INFO nova.compute.manager [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Terminating instance [ 1173.525062] env[62952]: DEBUG nova.compute.manager [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1173.525265] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1173.526118] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad8bcfb-f511-4cc1-86a1-c76075000d9b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.533671] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1173.533853] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b88f74a9-5fd3-4a21-9e84-2570e63f9bd1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.540142] env[62952]: DEBUG oslo_vmware.api [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1173.540142] env[62952]: value = "task-1263966" [ 1173.540142] env[62952]: _type = "Task" [ 1173.540142] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.547488] env[62952]: DEBUG oslo_vmware.api [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263966, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.819062] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263965, 'name': ReconfigVM_Task, 'duration_secs': 0.133215} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.819905] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea93401f-8faf-46b1-9738-6a23b158e759 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.837333] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1173.837584] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1173.837745] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1173.837932] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1173.838096] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1173.838259] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1173.838468] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1173.838629] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1173.838803] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1173.838963] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1173.839159] env[62952]: DEBUG nova.virt.hardware [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1173.839911] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ebf7e9f-c32d-41dc-a779-b9049d2562d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.842376] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 2b11ad9b-63d8-40da-8f68-3d8260c63bc7] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1173.846953] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1173.846953] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52dc66de-9ad1-d845-4754-557daa04e7e1" [ 1173.846953] env[62952]: _type = "Task" [ 1173.846953] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.854557] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52dc66de-9ad1-d845-4754-557daa04e7e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.932491] env[62952]: DEBUG nova.compute.manager [None req-57119c01-9b76-4012-9126-cc1162ac0eef tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Found 1 images (rotation: 2) {{(pid=62952) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1174.050664] env[62952]: DEBUG oslo_vmware.api [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263966, 'name': PowerOffVM_Task, 'duration_secs': 0.224981} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.050931] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1174.051121] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1174.051375] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f5e51e3-d4d8-4c60-b408-63299299f369 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.118451] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1174.118691] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1174.118885] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Deleting the datastore file [datastore2] 42793459-89e6-4941-9b20-cbe52a241c6d {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1174.119172] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-56ea5683-80fb-416f-a5fa-c1a3d6f0614a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.126298] env[62952]: DEBUG oslo_vmware.api [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for the task: (returnval){ [ 1174.126298] env[62952]: value = "task-1263968" [ 1174.126298] env[62952]: _type = "Task" [ 1174.126298] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.133667] env[62952]: DEBUG oslo_vmware.api [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263968, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.346685] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 67b729be-fb66-4ed1-bbea-e62216d460d5] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1174.357967] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52dc66de-9ad1-d845-4754-557daa04e7e1, 'name': SearchDatastore_Task, 'duration_secs': 0.007396} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.364075] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1174.364367] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f16bd7f8-e0c1-4857-b2b3-ad2ec339ec42 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.383232] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1174.383232] env[62952]: value = "task-1263969" [ 1174.383232] env[62952]: _type = "Task" [ 1174.383232] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.391189] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263969, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.637152] env[62952]: DEBUG oslo_vmware.api [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Task: {'id': task-1263968, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132441} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.637423] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1174.637605] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1174.637777] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1174.637952] env[62952]: INFO nova.compute.manager [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1174.638215] env[62952]: DEBUG oslo.service.loopingcall [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1174.638724] env[62952]: DEBUG nova.compute.manager [-] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1174.638822] env[62952]: DEBUG nova.network.neutron [-] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1174.852626] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 0533ee81-b69d-4f4e-9832-78c5ab4512b1] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1174.896165] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263969, 'name': ReconfigVM_Task, 'duration_secs': 0.233422} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.896505] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1174.897434] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cadb72d-2a3e-4b70-931c-4ca012a5b970 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.925715] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 43d79d26-4350-4329-bffc-cedfada40ee0/43d79d26-4350-4329-bffc-cedfada40ee0.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1174.926891] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1205a1ae-400b-4657-93b4-9850cb293392 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.953146] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1174.953146] env[62952]: value = "task-1263970" [ 1174.953146] env[62952]: _type = "Task" [ 1174.953146] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.964665] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263970, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.107336] env[62952]: DEBUG nova.compute.manager [req-b1a33d40-3640-402d-a0a4-940802490048 req-7d0fab0a-6581-4395-bf3f-920ec8c86a33 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Received event network-vif-deleted-ddd93d49-eb57-4af9-a9bb-35d8269b939d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1175.107419] env[62952]: INFO nova.compute.manager [req-b1a33d40-3640-402d-a0a4-940802490048 req-7d0fab0a-6581-4395-bf3f-920ec8c86a33 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Neutron deleted interface ddd93d49-eb57-4af9-a9bb-35d8269b939d; detaching it from the instance and deleting it from the info cache [ 1175.107723] env[62952]: DEBUG nova.network.neutron [req-b1a33d40-3640-402d-a0a4-940802490048 req-7d0fab0a-6581-4395-bf3f-920ec8c86a33 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1175.246533] env[62952]: DEBUG nova.compute.manager [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1175.247586] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4234e0b-c40f-40fb-8b20-97d3e80ab25c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.356047] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: f906dc47-12cb-46f5-ae5c-53cfcc8765a8] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1175.462309] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263970, 'name': ReconfigVM_Task, 'duration_secs': 0.26849} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.462597] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 43d79d26-4350-4329-bffc-cedfada40ee0/43d79d26-4350-4329-bffc-cedfada40ee0.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1175.463564] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5014c19-5c55-4bba-990d-c5d8dce6d35a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.482476] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3668b667-d928-473c-86ca-48191c34a7a6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.499597] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3583272-826c-4bfe-a8d8-9e6f4dbb4f27 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.516487] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a1d03a-bb79-44df-9e5b-f2fea6009b06 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.522540] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1175.522748] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7387b5a4-3b42-4247-aea3-341b75bcdb10 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.528517] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1175.528517] env[62952]: value = "task-1263971" [ 1175.528517] env[62952]: _type = "Task" [ 1175.528517] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.535262] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263971, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.580842] env[62952]: DEBUG nova.network.neutron [-] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1175.611293] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-94fdd413-e40d-46c3-82d9-9d4120c94ce0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.620708] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4a3859-3a40-4e85-adf5-ff0d1a6a1130 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.644967] env[62952]: DEBUG nova.compute.manager [req-b1a33d40-3640-402d-a0a4-940802490048 req-7d0fab0a-6581-4395-bf3f-920ec8c86a33 service nova] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Detach interface failed, port_id=ddd93d49-eb57-4af9-a9bb-35d8269b939d, reason: Instance 42793459-89e6-4941-9b20-cbe52a241c6d could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1175.759170] env[62952]: INFO nova.compute.manager [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] instance snapshotting [ 1175.759933] env[62952]: DEBUG nova.objects.instance [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lazy-loading 'flavor' on Instance uuid f92ef782-fd64-47f0-8145-9261a1c3005c {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1175.859284] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 58e703ea-7ccf-4114-a5a2-49533390ed37] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1176.038449] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263971, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.084105] env[62952]: INFO nova.compute.manager [-] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Took 1.44 seconds to deallocate network for instance. [ 1176.267187] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2836e3e-4651-41e4-8659-77181d95085e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.286083] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6180c6a-b565-4840-867d-6c614ca21a9f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.361971] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: c41f8335-cb2d-4d26-9ae3-2ccf0ee66adb] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1176.540063] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263971, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.591832] env[62952]: DEBUG oslo_concurrency.lockutils [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.591832] env[62952]: DEBUG oslo_concurrency.lockutils [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.591832] env[62952]: DEBUG nova.objects.instance [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lazy-loading 'resources' on Instance uuid 42793459-89e6-4941-9b20-cbe52a241c6d {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1176.796104] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1176.796422] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-75efade6-bf2f-4197-a227-e23bf836c367 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.804697] env[62952]: DEBUG oslo_vmware.api [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1176.804697] env[62952]: value = "task-1263972" [ 1176.804697] env[62952]: _type = "Task" [ 1176.804697] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.812714] env[62952]: DEBUG oslo_vmware.api [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263972, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.865899] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: bc369b94-2f40-444e-88d7-069fd33e2646] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1177.039295] env[62952]: DEBUG oslo_vmware.api [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263971, 'name': PowerOnVM_Task, 'duration_secs': 1.169442} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.039581] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1177.144697] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2fd91a-ce60-4d9f-8818-686531599c53 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.151740] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e33bfde-097e-4053-b90d-02cfa0033941 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.182080] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1b4d918-47ba-4dde-b7a7-d16c1b4a9d9e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.189821] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13836c9d-fb94-45a7-a7e8-2b0ae6e2f4e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.203102] env[62952]: DEBUG nova.compute.provider_tree [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1177.314336] env[62952]: DEBUG oslo_vmware.api [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263972, 'name': CreateSnapshot_Task, 'duration_secs': 0.405947} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.314593] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1177.315320] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aa412e0-750c-4afb-a043-6291a0853cb4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.368653] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 219cfb14-ca73-45d3-86e8-ed4642d7d480] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1177.706140] env[62952]: DEBUG nova.scheduler.client.report [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1177.832310] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1177.832619] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5607a64e-8c92-4fe9-90d9-c7624d05ae3d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.840260] env[62952]: DEBUG oslo_vmware.api [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1177.840260] env[62952]: value = "task-1263973" [ 1177.840260] env[62952]: _type = "Task" [ 1177.840260] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.848153] env[62952]: DEBUG oslo_vmware.api [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263973, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.871977] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 9cc7bc2f-ff22-43cf-be79-ff8c6426115b] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1178.050596] env[62952]: INFO nova.compute.manager [None req-17b85bc6-c3df-4b82-9362-84e213af4df7 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating instance to original state: 'active' [ 1178.211148] env[62952]: DEBUG oslo_concurrency.lockutils [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.619s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.229607] env[62952]: INFO nova.scheduler.client.report [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Deleted allocations for instance 42793459-89e6-4941-9b20-cbe52a241c6d [ 1178.351250] env[62952]: DEBUG oslo_vmware.api [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263973, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.375325] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 5f1eca6f-4cc4-483e-b72a-b479378277fe] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1178.736954] env[62952]: DEBUG oslo_concurrency.lockutils [None req-577640d3-1957-465a-abd9-fbfa309da433 tempest-AttachVolumeShelveTestJSON-303067515 tempest-AttachVolumeShelveTestJSON-303067515-project-member] Lock "42793459-89e6-4941-9b20-cbe52a241c6d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.216s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.851796] env[62952]: DEBUG oslo_vmware.api [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263973, 'name': CloneVM_Task, 'duration_secs': 0.95523} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.852334] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Created linked-clone VM from snapshot [ 1178.853285] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373d612e-d6d1-4dbc-b2a3-08cdea0f4d4f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.862409] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Uploading image 06645474-95de-4f92-9820-3554090526cd {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1178.880574] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: f4f8bee4-bf92-4321-afdc-16f7fb17dbae] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1178.893300] env[62952]: DEBUG oslo_vmware.rw_handles [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1178.893300] env[62952]: value = "vm-271987" [ 1178.893300] env[62952]: _type = "VirtualMachine" [ 1178.893300] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1178.894358] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-82560695-c887-4115-9b71-7609242cdb31 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.901819] env[62952]: DEBUG oslo_vmware.rw_handles [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lease: (returnval){ [ 1178.901819] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c8a5e9-db3c-f87e-7d0a-619c37207b8e" [ 1178.901819] env[62952]: _type = "HttpNfcLease" [ 1178.901819] env[62952]: } obtained for exporting VM: (result){ [ 1178.901819] env[62952]: value = "vm-271987" [ 1178.901819] env[62952]: _type = "VirtualMachine" [ 1178.901819] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1178.902151] env[62952]: DEBUG oslo_vmware.api [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the lease: (returnval){ [ 1178.902151] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c8a5e9-db3c-f87e-7d0a-619c37207b8e" [ 1178.902151] env[62952]: _type = "HttpNfcLease" [ 1178.902151] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1178.910575] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1178.910575] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c8a5e9-db3c-f87e-7d0a-619c37207b8e" [ 1178.910575] env[62952]: _type = "HttpNfcLease" [ 1178.910575] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1179.374056] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "43d79d26-4350-4329-bffc-cedfada40ee0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1179.374347] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "43d79d26-4350-4329-bffc-cedfada40ee0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1179.374567] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "43d79d26-4350-4329-bffc-cedfada40ee0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1179.374754] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "43d79d26-4350-4329-bffc-cedfada40ee0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1179.374927] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "43d79d26-4350-4329-bffc-cedfada40ee0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.377130] env[62952]: INFO nova.compute.manager [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Terminating instance [ 1179.379331] env[62952]: DEBUG nova.compute.manager [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1179.379562] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1179.380367] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8212956-73eb-4aa4-8a9d-34074bc934c7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.384677] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1179.387897] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1179.388142] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c93c3248-f3a5-4c93-93bd-eb87f2b6ab7f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.394646] env[62952]: DEBUG oslo_vmware.api [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1179.394646] env[62952]: value = "task-1263975" [ 1179.394646] env[62952]: _type = "Task" [ 1179.394646] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.402791] env[62952]: DEBUG oslo_vmware.api [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263975, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.409352] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1179.409352] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c8a5e9-db3c-f87e-7d0a-619c37207b8e" [ 1179.409352] env[62952]: _type = "HttpNfcLease" [ 1179.409352] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1179.409660] env[62952]: DEBUG oslo_vmware.rw_handles [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1179.409660] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c8a5e9-db3c-f87e-7d0a-619c37207b8e" [ 1179.409660] env[62952]: _type = "HttpNfcLease" [ 1179.409660] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1179.410389] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-919a6c7b-3456-4a11-b0e5-605221e4279c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.417835] env[62952]: DEBUG oslo_vmware.rw_handles [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f472b3-0098-5ab5-f8bd-46e45a562291/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1179.418041] env[62952]: DEBUG oslo_vmware.rw_handles [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f472b3-0098-5ab5-f8bd-46e45a562291/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1179.507455] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-35845bbf-d65e-477e-af32-04cb7e64b2ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.906228] env[62952]: DEBUG oslo_vmware.api [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263975, 'name': PowerOffVM_Task, 'duration_secs': 0.213984} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.906730] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1179.907023] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1179.907766] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3c321b76-da0a-4388-8c2f-624056518b1c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.969676] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1179.969855] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1179.970182] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Deleting the datastore file [datastore1] 43d79d26-4350-4329-bffc-cedfada40ee0 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1179.970534] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-921c950d-d9f1-4088-a616-06ad89321108 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.977435] env[62952]: DEBUG oslo_vmware.api [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1179.977435] env[62952]: value = "task-1263977" [ 1179.977435] env[62952]: _type = "Task" [ 1179.977435] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.988078] env[62952]: DEBUG oslo_vmware.api [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263977, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.383810] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.384210] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.487435] env[62952]: DEBUG oslo_vmware.api [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263977, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165864} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.487798] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1180.488072] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1180.488311] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1180.488507] env[62952]: INFO nova.compute.manager [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1180.488802] env[62952]: DEBUG oslo.service.loopingcall [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1180.489048] env[62952]: DEBUG nova.compute.manager [-] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1180.489220] env[62952]: DEBUG nova.network.neutron [-] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1180.891544] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.891544] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1180.891544] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1180.948718] env[62952]: DEBUG nova.compute.manager [req-b85ce8ca-fb29-42ff-a8f8-3a97a5d6865c req-c1479e91-2f78-4f28-bb17-e0417ecc1855 service nova] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Received event network-vif-deleted-e6fdb64f-1d51-4b10-b297-9f760ae81c95 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1180.948963] env[62952]: INFO nova.compute.manager [req-b85ce8ca-fb29-42ff-a8f8-3a97a5d6865c req-c1479e91-2f78-4f28-bb17-e0417ecc1855 service nova] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Neutron deleted interface e6fdb64f-1d51-4b10-b297-9f760ae81c95; detaching it from the instance and deleting it from the info cache [ 1180.949244] env[62952]: DEBUG nova.network.neutron [req-b85ce8ca-fb29-42ff-a8f8-3a97a5d6865c req-c1479e91-2f78-4f28-bb17-e0417ecc1855 service nova] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1181.393994] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Skipping network cache update for instance because it is being deleted. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 1181.394430] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1181.394430] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquired lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.394536] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Forcefully refreshing network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1181.394643] env[62952]: DEBUG nova.objects.instance [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lazy-loading 'info_cache' on Instance uuid f92ef782-fd64-47f0-8145-9261a1c3005c {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1181.421752] env[62952]: DEBUG nova.network.neutron [-] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1181.452374] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d490a3b6-f2fe-4726-ae62-912ac80e4808 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.462945] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af4f340-7103-4180-a516-b6929d02cac9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.486308] env[62952]: DEBUG nova.compute.manager [req-b85ce8ca-fb29-42ff-a8f8-3a97a5d6865c req-c1479e91-2f78-4f28-bb17-e0417ecc1855 service nova] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Detach interface failed, port_id=e6fdb64f-1d51-4b10-b297-9f760ae81c95, reason: Instance 43d79d26-4350-4329-bffc-cedfada40ee0 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1181.924431] env[62952]: INFO nova.compute.manager [-] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Took 1.43 seconds to deallocate network for instance. [ 1182.431303] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.431583] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.431783] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1182.457953] env[62952]: INFO nova.scheduler.client.report [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Deleted allocations for instance 43d79d26-4350-4329-bffc-cedfada40ee0 [ 1182.965561] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bee6d099-7683-4189-bbfd-815167cefc18 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "43d79d26-4350-4329-bffc-cedfada40ee0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.591s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.288681] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance_info_cache with network_info: [{"id": "19fcce34-055a-492e-8c73-a39767f93814", "address": "fa:16:3e:71:3d:5f", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19fcce34-05", "ovs_interfaceid": "19fcce34-055a-492e-8c73-a39767f93814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1183.791371] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Releasing lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.791686] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updated the network info_cache for instance {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1183.791835] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1183.792082] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1183.792214] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1183.792357] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1183.792597] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1183.792692] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1184.481071] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.481247] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.984516] env[62952]: DEBUG nova.compute.manager [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1185.509807] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.509959] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.511546] env[62952]: INFO nova.compute.claims [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1186.570036] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-784b041c-45b6-44a9-9555-11d757e36100 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.578640] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1050294a-8781-498d-b065-a5c2e84c4c52 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.610963] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b0eb35-1e28-4de3-b523-7c5662a2c638 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.618534] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a60a389-0f5c-43f5-a081-e3db0167a28b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.633750] env[62952]: DEBUG nova.compute.provider_tree [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1187.137071] env[62952]: DEBUG nova.scheduler.client.report [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1187.636587] env[62952]: DEBUG oslo_vmware.rw_handles [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f472b3-0098-5ab5-f8bd-46e45a562291/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1187.637563] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba291bf-9f9b-4d74-8e86-0b8f83a422e0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.644037] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.134s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.644539] env[62952]: DEBUG nova.compute.manager [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1187.647175] env[62952]: DEBUG oslo_vmware.rw_handles [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f472b3-0098-5ab5-f8bd-46e45a562291/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1187.647356] env[62952]: ERROR oslo_vmware.rw_handles [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f472b3-0098-5ab5-f8bd-46e45a562291/disk-0.vmdk due to incomplete transfer. [ 1187.647817] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-df5bbf2b-ddb5-4468-9be0-c0be3297e497 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.655375] env[62952]: DEBUG oslo_vmware.rw_handles [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f472b3-0098-5ab5-f8bd-46e45a562291/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1187.655617] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Uploaded image 06645474-95de-4f92-9820-3554090526cd to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1187.658016] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1187.658295] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-343c82dc-772b-486b-9c8f-d472af6220da {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.663504] env[62952]: DEBUG oslo_vmware.api [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1187.663504] env[62952]: value = "task-1263979" [ 1187.663504] env[62952]: _type = "Task" [ 1187.663504] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.672381] env[62952]: DEBUG oslo_vmware.api [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263979, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.150481] env[62952]: DEBUG nova.compute.utils [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1188.152226] env[62952]: DEBUG nova.compute.manager [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1188.152226] env[62952]: DEBUG nova.network.neutron [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1188.174492] env[62952]: DEBUG oslo_vmware.api [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263979, 'name': Destroy_Task, 'duration_secs': 0.303736} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.174779] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Destroyed the VM [ 1188.175018] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1188.175308] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c577ad4f-3c4d-4272-bcfe-99d14a70c329 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.182755] env[62952]: DEBUG oslo_vmware.api [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1188.182755] env[62952]: value = "task-1263980" [ 1188.182755] env[62952]: _type = "Task" [ 1188.182755] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.190912] env[62952]: DEBUG oslo_vmware.api [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263980, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.197844] env[62952]: DEBUG nova.policy [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ffc122d64b19432aae8d9627d2928c06', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5080fbeac0744478843ab28efea8fc18', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1188.211788] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "c5d21282-1a4b-4998-8170-ee5423b958ad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.212019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "c5d21282-1a4b-4998-8170-ee5423b958ad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.483237] env[62952]: DEBUG nova.network.neutron [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Successfully created port: c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1188.655571] env[62952]: DEBUG nova.compute.manager [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1188.693700] env[62952]: DEBUG oslo_vmware.api [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263980, 'name': RemoveSnapshot_Task} progress is 17%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.714337] env[62952]: DEBUG nova.compute.manager [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1189.193826] env[62952]: DEBUG oslo_vmware.api [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263980, 'name': RemoveSnapshot_Task, 'duration_secs': 0.622323} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.194173] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1189.194412] env[62952]: INFO nova.compute.manager [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Took 12.93 seconds to snapshot the instance on the hypervisor. [ 1189.240970] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1189.241253] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.242808] env[62952]: INFO nova.compute.claims [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1189.664708] env[62952]: DEBUG nova.compute.manager [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1189.690810] env[62952]: DEBUG nova.virt.hardware [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1189.691089] env[62952]: DEBUG nova.virt.hardware [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1189.691285] env[62952]: DEBUG nova.virt.hardware [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1189.691567] env[62952]: DEBUG nova.virt.hardware [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1189.691736] env[62952]: DEBUG nova.virt.hardware [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1189.691912] env[62952]: DEBUG nova.virt.hardware [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1189.692180] env[62952]: DEBUG nova.virt.hardware [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1189.692367] env[62952]: DEBUG nova.virt.hardware [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1189.692542] env[62952]: DEBUG nova.virt.hardware [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1189.692809] env[62952]: DEBUG nova.virt.hardware [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1189.692881] env[62952]: DEBUG nova.virt.hardware [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1189.693761] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6951c57e-4623-4efa-bfb4-24d7b062e1b9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.704935] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90bcbdf7-be65-4bd0-9d85-61048a4b7596 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.743943] env[62952]: DEBUG nova.compute.manager [None req-12e2f7ed-0361-4308-ba9b-c685ecab0774 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Found 2 images (rotation: 2) {{(pid=62952) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1189.863034] env[62952]: DEBUG nova.compute.manager [req-c914a57f-e0b3-4263-84cf-390d3ec8063d req-c327f078-2803-4c7b-974e-e6f43551bc8b service nova] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Received event network-vif-plugged-c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1189.863355] env[62952]: DEBUG oslo_concurrency.lockutils [req-c914a57f-e0b3-4263-84cf-390d3ec8063d req-c327f078-2803-4c7b-974e-e6f43551bc8b service nova] Acquiring lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1189.863583] env[62952]: DEBUG oslo_concurrency.lockutils [req-c914a57f-e0b3-4263-84cf-390d3ec8063d req-c327f078-2803-4c7b-974e-e6f43551bc8b service nova] Lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.863798] env[62952]: DEBUG oslo_concurrency.lockutils [req-c914a57f-e0b3-4263-84cf-390d3ec8063d req-c327f078-2803-4c7b-974e-e6f43551bc8b service nova] Lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.864058] env[62952]: DEBUG nova.compute.manager [req-c914a57f-e0b3-4263-84cf-390d3ec8063d req-c327f078-2803-4c7b-974e-e6f43551bc8b service nova] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] No waiting events found dispatching network-vif-plugged-c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1189.864293] env[62952]: WARNING nova.compute.manager [req-c914a57f-e0b3-4263-84cf-390d3ec8063d req-c327f078-2803-4c7b-974e-e6f43551bc8b service nova] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Received unexpected event network-vif-plugged-c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8 for instance with vm_state building and task_state spawning. [ 1189.949464] env[62952]: DEBUG nova.network.neutron [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Successfully updated port: c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1190.299598] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c64dfd-2a2f-45bd-9f90-5e5b2648d453 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.307238] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e437e1fc-440f-4b3c-97b6-4bda0f4404e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.337438] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c177c16b-5596-455b-9919-0dafcb2fc3fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.344839] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a66dbdf2-788f-492c-a1c8-18d396c060fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.357976] env[62952]: DEBUG nova.compute.provider_tree [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1190.452537] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "refresh_cache-b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1190.452537] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "refresh_cache-b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1190.452537] env[62952]: DEBUG nova.network.neutron [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1190.861403] env[62952]: DEBUG nova.scheduler.client.report [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1190.981942] env[62952]: DEBUG nova.network.neutron [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1191.112764] env[62952]: DEBUG nova.network.neutron [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Updating instance_info_cache with network_info: [{"id": "c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8", "address": "fa:16:3e:b0:e3:32", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3a25e8c-b6", "ovs_interfaceid": "c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1191.365681] env[62952]: DEBUG nova.compute.manager [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1191.366475] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.125s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.366946] env[62952]: DEBUG nova.compute.manager [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1191.370138] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159d5d75-d9dd-48be-812e-bfc4fd945278 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.615195] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "refresh_cache-b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1191.615542] env[62952]: DEBUG nova.compute.manager [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Instance network_info: |[{"id": "c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8", "address": "fa:16:3e:b0:e3:32", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3a25e8c-b6", "ovs_interfaceid": "c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1191.615990] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:e3:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e1c9bb98-73a9-48eb-856e-a541afe9b07b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1191.623320] env[62952]: DEBUG oslo.service.loopingcall [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1191.623536] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1191.623757] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1492cefe-1c5f-40de-9af4-5cf1dbf11e6d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.643013] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1191.643013] env[62952]: value = "task-1263981" [ 1191.643013] env[62952]: _type = "Task" [ 1191.643013] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.650094] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263981, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.875389] env[62952]: DEBUG nova.compute.utils [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1191.876892] env[62952]: DEBUG nova.compute.manager [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1191.877147] env[62952]: DEBUG nova.network.neutron [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1191.879719] env[62952]: INFO nova.compute.manager [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] instance snapshotting [ 1191.880308] env[62952]: DEBUG nova.objects.instance [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lazy-loading 'flavor' on Instance uuid f92ef782-fd64-47f0-8145-9261a1c3005c {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1191.888233] env[62952]: DEBUG nova.compute.manager [req-d116b397-0498-4eae-9189-2666a2dae933 req-440f5401-dec9-40c8-b0f0-bf0305beb497 service nova] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Received event network-changed-c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1191.888421] env[62952]: DEBUG nova.compute.manager [req-d116b397-0498-4eae-9189-2666a2dae933 req-440f5401-dec9-40c8-b0f0-bf0305beb497 service nova] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Refreshing instance network info cache due to event network-changed-c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1191.888767] env[62952]: DEBUG oslo_concurrency.lockutils [req-d116b397-0498-4eae-9189-2666a2dae933 req-440f5401-dec9-40c8-b0f0-bf0305beb497 service nova] Acquiring lock "refresh_cache-b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1191.888849] env[62952]: DEBUG oslo_concurrency.lockutils [req-d116b397-0498-4eae-9189-2666a2dae933 req-440f5401-dec9-40c8-b0f0-bf0305beb497 service nova] Acquired lock "refresh_cache-b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.889011] env[62952]: DEBUG nova.network.neutron [req-d116b397-0498-4eae-9189-2666a2dae933 req-440f5401-dec9-40c8-b0f0-bf0305beb497 service nova] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Refreshing network info cache for port c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1191.920623] env[62952]: DEBUG nova.policy [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0a36328ec404c22adf6d2d256c9514c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f725be1feb8a44478d297b0be58c59ba', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1192.154039] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263981, 'name': CreateVM_Task, 'duration_secs': 0.295595} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.154039] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1192.155069] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1192.155069] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.155069] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1192.155245] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-247dd784-7204-448e-97eb-e2c4de162bc5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.159682] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1192.159682] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d2adda-21da-fa3e-bc14-54c8db996f7e" [ 1192.159682] env[62952]: _type = "Task" [ 1192.159682] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.167828] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d2adda-21da-fa3e-bc14-54c8db996f7e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.199837] env[62952]: DEBUG nova.network.neutron [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Successfully created port: f5eb1e66-2290-4179-b630-173dd8f3f058 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1192.380826] env[62952]: DEBUG nova.compute.manager [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1192.386025] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb03497-e565-4574-a5a0-17352a48dfa3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.406633] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a97c58-149f-4f75-8426-6d1dd21f7526 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.602539] env[62952]: DEBUG nova.network.neutron [req-d116b397-0498-4eae-9189-2666a2dae933 req-440f5401-dec9-40c8-b0f0-bf0305beb497 service nova] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Updated VIF entry in instance network info cache for port c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1192.602899] env[62952]: DEBUG nova.network.neutron [req-d116b397-0498-4eae-9189-2666a2dae933 req-440f5401-dec9-40c8-b0f0-bf0305beb497 service nova] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Updating instance_info_cache with network_info: [{"id": "c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8", "address": "fa:16:3e:b0:e3:32", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3a25e8c-b6", "ovs_interfaceid": "c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1192.669741] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d2adda-21da-fa3e-bc14-54c8db996f7e, 'name': SearchDatastore_Task, 'duration_secs': 0.010491} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.670122] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1192.670263] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1192.670499] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1192.670646] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.670824] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1192.671164] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-860ea822-0745-4891-b88f-976614677760 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.679296] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1192.679468] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1192.680158] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77603a6f-b9db-454a-82fd-3d0304b641de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.684903] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1192.684903] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52053cb6-1a28-7bcc-40fb-069c4f15b7ce" [ 1192.684903] env[62952]: _type = "Task" [ 1192.684903] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.691959] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52053cb6-1a28-7bcc-40fb-069c4f15b7ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.917444] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1192.917915] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-382500ef-ecb7-4222-b1b2-f0975a97b351 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.926841] env[62952]: DEBUG oslo_vmware.api [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1192.926841] env[62952]: value = "task-1263982" [ 1192.926841] env[62952]: _type = "Task" [ 1192.926841] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.934714] env[62952]: DEBUG oslo_vmware.api [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263982, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.106032] env[62952]: DEBUG oslo_concurrency.lockutils [req-d116b397-0498-4eae-9189-2666a2dae933 req-440f5401-dec9-40c8-b0f0-bf0305beb497 service nova] Releasing lock "refresh_cache-b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1193.196587] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52053cb6-1a28-7bcc-40fb-069c4f15b7ce, 'name': SearchDatastore_Task, 'duration_secs': 0.007848} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.197349] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54e744e9-989d-4666-9678-8b705754b7eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.202557] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1193.202557] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e5ebec-1b96-1006-ac71-57f7aa6d32e5" [ 1193.202557] env[62952]: _type = "Task" [ 1193.202557] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.211734] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e5ebec-1b96-1006-ac71-57f7aa6d32e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.392418] env[62952]: DEBUG nova.compute.manager [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1193.417754] env[62952]: DEBUG nova.virt.hardware [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1193.418059] env[62952]: DEBUG nova.virt.hardware [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1193.418230] env[62952]: DEBUG nova.virt.hardware [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1193.418413] env[62952]: DEBUG nova.virt.hardware [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1193.418560] env[62952]: DEBUG nova.virt.hardware [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1193.418707] env[62952]: DEBUG nova.virt.hardware [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1193.418910] env[62952]: DEBUG nova.virt.hardware [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1193.419112] env[62952]: DEBUG nova.virt.hardware [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1193.419314] env[62952]: DEBUG nova.virt.hardware [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1193.419482] env[62952]: DEBUG nova.virt.hardware [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1193.419656] env[62952]: DEBUG nova.virt.hardware [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1193.420529] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61961b39-a5ce-4819-9481-4e9b66297500 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.111947] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c936ef-7553-4d18-8238-fd92764bc769 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.116338] env[62952]: DEBUG nova.network.neutron [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Successfully updated port: f5eb1e66-2290-4179-b630-173dd8f3f058 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1194.119412] env[62952]: DEBUG nova.compute.manager [req-4e2c792f-d7a7-47a3-9b83-a91b8f67fdb6 req-1ad275c0-f631-4e8c-b6e0-c93730c12a2d service nova] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Received event network-vif-plugged-f5eb1e66-2290-4179-b630-173dd8f3f058 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1194.119648] env[62952]: DEBUG oslo_concurrency.lockutils [req-4e2c792f-d7a7-47a3-9b83-a91b8f67fdb6 req-1ad275c0-f631-4e8c-b6e0-c93730c12a2d service nova] Acquiring lock "c5d21282-1a4b-4998-8170-ee5423b958ad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.119865] env[62952]: DEBUG oslo_concurrency.lockutils [req-4e2c792f-d7a7-47a3-9b83-a91b8f67fdb6 req-1ad275c0-f631-4e8c-b6e0-c93730c12a2d service nova] Lock "c5d21282-1a4b-4998-8170-ee5423b958ad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1194.120045] env[62952]: DEBUG oslo_concurrency.lockutils [req-4e2c792f-d7a7-47a3-9b83-a91b8f67fdb6 req-1ad275c0-f631-4e8c-b6e0-c93730c12a2d service nova] Lock "c5d21282-1a4b-4998-8170-ee5423b958ad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.120218] env[62952]: DEBUG nova.compute.manager [req-4e2c792f-d7a7-47a3-9b83-a91b8f67fdb6 req-1ad275c0-f631-4e8c-b6e0-c93730c12a2d service nova] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] No waiting events found dispatching network-vif-plugged-f5eb1e66-2290-4179-b630-173dd8f3f058 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1194.120384] env[62952]: WARNING nova.compute.manager [req-4e2c792f-d7a7-47a3-9b83-a91b8f67fdb6 req-1ad275c0-f631-4e8c-b6e0-c93730c12a2d service nova] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Received unexpected event network-vif-plugged-f5eb1e66-2290-4179-b630-173dd8f3f058 for instance with vm_state building and task_state spawning. [ 1194.120541] env[62952]: DEBUG nova.compute.manager [req-4e2c792f-d7a7-47a3-9b83-a91b8f67fdb6 req-1ad275c0-f631-4e8c-b6e0-c93730c12a2d service nova] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Received event network-changed-f5eb1e66-2290-4179-b630-173dd8f3f058 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1194.120691] env[62952]: DEBUG nova.compute.manager [req-4e2c792f-d7a7-47a3-9b83-a91b8f67fdb6 req-1ad275c0-f631-4e8c-b6e0-c93730c12a2d service nova] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Refreshing instance network info cache due to event network-changed-f5eb1e66-2290-4179-b630-173dd8f3f058. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1194.120870] env[62952]: DEBUG oslo_concurrency.lockutils [req-4e2c792f-d7a7-47a3-9b83-a91b8f67fdb6 req-1ad275c0-f631-4e8c-b6e0-c93730c12a2d service nova] Acquiring lock "refresh_cache-c5d21282-1a4b-4998-8170-ee5423b958ad" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1194.121047] env[62952]: DEBUG oslo_concurrency.lockutils [req-4e2c792f-d7a7-47a3-9b83-a91b8f67fdb6 req-1ad275c0-f631-4e8c-b6e0-c93730c12a2d service nova] Acquired lock "refresh_cache-c5d21282-1a4b-4998-8170-ee5423b958ad" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1194.121219] env[62952]: DEBUG nova.network.neutron [req-4e2c792f-d7a7-47a3-9b83-a91b8f67fdb6 req-1ad275c0-f631-4e8c-b6e0-c93730c12a2d service nova] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Refreshing network info cache for port f5eb1e66-2290-4179-b630-173dd8f3f058 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1194.131135] env[62952]: DEBUG oslo_vmware.api [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263982, 'name': CreateSnapshot_Task, 'duration_secs': 0.39704} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.142324] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1194.142600] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e5ebec-1b96-1006-ac71-57f7aa6d32e5, 'name': SearchDatastore_Task, 'duration_secs': 0.009146} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.143842] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c9bd5ca-e352-46bc-b416-4762056d6e2a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.146172] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1194.146424] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] b01ea5c5-1d49-47e7-ada2-acbacbe5abdb/b01ea5c5-1d49-47e7-ada2-acbacbe5abdb.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1194.146871] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f7d9b2fc-bb48-4a68-bef7-810b3dfc545c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.156468] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1194.156468] env[62952]: value = "task-1263983" [ 1194.156468] env[62952]: _type = "Task" [ 1194.156468] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.163893] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263983, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.664513] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1194.665716] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "refresh_cache-c5d21282-1a4b-4998-8170-ee5423b958ad" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1194.665969] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9b4f3994-1e31-4065-a2de-51f1c0e3e3ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.677577] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263983, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.411229} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.678665] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] b01ea5c5-1d49-47e7-ada2-acbacbe5abdb/b01ea5c5-1d49-47e7-ada2-acbacbe5abdb.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1194.678888] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1194.679205] env[62952]: DEBUG oslo_vmware.api [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1194.679205] env[62952]: value = "task-1263984" [ 1194.679205] env[62952]: _type = "Task" [ 1194.679205] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.679413] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-06a4a747-34db-4520-a54c-8f848eb88fbf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.688674] env[62952]: DEBUG oslo_vmware.api [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263984, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.689708] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1194.689708] env[62952]: value = "task-1263985" [ 1194.689708] env[62952]: _type = "Task" [ 1194.689708] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.696527] env[62952]: DEBUG nova.network.neutron [req-4e2c792f-d7a7-47a3-9b83-a91b8f67fdb6 req-1ad275c0-f631-4e8c-b6e0-c93730c12a2d service nova] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1194.701166] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263985, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.770210] env[62952]: DEBUG nova.network.neutron [req-4e2c792f-d7a7-47a3-9b83-a91b8f67fdb6 req-1ad275c0-f631-4e8c-b6e0-c93730c12a2d service nova] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1195.191912] env[62952]: DEBUG oslo_vmware.api [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263984, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.199245] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263985, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058008} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.199509] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1195.200298] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c6af98-fe10-4f18-94cd-80215ca4085c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.221342] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] b01ea5c5-1d49-47e7-ada2-acbacbe5abdb/b01ea5c5-1d49-47e7-ada2-acbacbe5abdb.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1195.221605] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d0d7171-886e-46bb-8ef7-39d3bcdd0dc2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.240396] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1195.240396] env[62952]: value = "task-1263986" [ 1195.240396] env[62952]: _type = "Task" [ 1195.240396] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.247974] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263986, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.273099] env[62952]: DEBUG oslo_concurrency.lockutils [req-4e2c792f-d7a7-47a3-9b83-a91b8f67fdb6 req-1ad275c0-f631-4e8c-b6e0-c93730c12a2d service nova] Releasing lock "refresh_cache-c5d21282-1a4b-4998-8170-ee5423b958ad" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1195.273447] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquired lock "refresh_cache-c5d21282-1a4b-4998-8170-ee5423b958ad" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1195.273605] env[62952]: DEBUG nova.network.neutron [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1195.691450] env[62952]: DEBUG oslo_vmware.api [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263984, 'name': CloneVM_Task} progress is 95%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.749699] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263986, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.804734] env[62952]: DEBUG nova.network.neutron [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1195.926710] env[62952]: DEBUG nova.network.neutron [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Updating instance_info_cache with network_info: [{"id": "f5eb1e66-2290-4179-b630-173dd8f3f058", "address": "fa:16:3e:15:7b:1a", "network": {"id": "8e167a8b-2628-43e1-8418-e673064745b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-232555460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f725be1feb8a44478d297b0be58c59ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5eb1e66-22", "ovs_interfaceid": "f5eb1e66-2290-4179-b630-173dd8f3f058", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1196.193095] env[62952]: DEBUG oslo_vmware.api [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263984, 'name': CloneVM_Task, 'duration_secs': 1.214196} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.193095] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Created linked-clone VM from snapshot [ 1196.193753] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc7a5cdf-0852-4593-a9af-677484cc0799 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.200902] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Uploading image 4bad21a1-8a67-4b7d-acb8-57b395177cc7 {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1196.221076] env[62952]: DEBUG oslo_vmware.rw_handles [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1196.221076] env[62952]: value = "vm-271990" [ 1196.221076] env[62952]: _type = "VirtualMachine" [ 1196.221076] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1196.221335] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-cf3a247d-c8ea-4742-aa35-fd66e0b0a2e3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.229183] env[62952]: DEBUG oslo_vmware.rw_handles [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lease: (returnval){ [ 1196.229183] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ebc440-49a5-10eb-9377-e367ac2d6dea" [ 1196.229183] env[62952]: _type = "HttpNfcLease" [ 1196.229183] env[62952]: } obtained for exporting VM: (result){ [ 1196.229183] env[62952]: value = "vm-271990" [ 1196.229183] env[62952]: _type = "VirtualMachine" [ 1196.229183] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1196.229461] env[62952]: DEBUG oslo_vmware.api [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the lease: (returnval){ [ 1196.229461] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ebc440-49a5-10eb-9377-e367ac2d6dea" [ 1196.229461] env[62952]: _type = "HttpNfcLease" [ 1196.229461] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1196.235055] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1196.235055] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ebc440-49a5-10eb-9377-e367ac2d6dea" [ 1196.235055] env[62952]: _type = "HttpNfcLease" [ 1196.235055] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1196.248531] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263986, 'name': ReconfigVM_Task, 'duration_secs': 0.528581} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.248713] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Reconfigured VM instance instance-0000006a to attach disk [datastore1] b01ea5c5-1d49-47e7-ada2-acbacbe5abdb/b01ea5c5-1d49-47e7-ada2-acbacbe5abdb.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1196.249313] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6b56eca6-571b-491a-8f14-3a52fa286a20 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.254896] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1196.254896] env[62952]: value = "task-1263988" [ 1196.254896] env[62952]: _type = "Task" [ 1196.254896] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.261862] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263988, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.429866] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Releasing lock "refresh_cache-c5d21282-1a4b-4998-8170-ee5423b958ad" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1196.430248] env[62952]: DEBUG nova.compute.manager [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Instance network_info: |[{"id": "f5eb1e66-2290-4179-b630-173dd8f3f058", "address": "fa:16:3e:15:7b:1a", "network": {"id": "8e167a8b-2628-43e1-8418-e673064745b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-232555460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f725be1feb8a44478d297b0be58c59ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5eb1e66-22", "ovs_interfaceid": "f5eb1e66-2290-4179-b630-173dd8f3f058", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1196.430670] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:7b:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69351262-8d39-441a-85ba-3a78df436d17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f5eb1e66-2290-4179-b630-173dd8f3f058', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1196.437867] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Creating folder: Project (f725be1feb8a44478d297b0be58c59ba). Parent ref: group-v271811. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1196.438160] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cf5d19aa-5c4f-40d6-971b-9bafded6f0b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.448547] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Created folder: Project (f725be1feb8a44478d297b0be58c59ba) in parent group-v271811. [ 1196.448694] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Creating folder: Instances. Parent ref: group-v271991. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1196.448921] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f129b4ed-0208-4423-9c7d-2d0d0e7fa0b0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.458765] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Created folder: Instances in parent group-v271991. [ 1196.458977] env[62952]: DEBUG oslo.service.loopingcall [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1196.459176] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1196.459367] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e34b8964-22aa-4099-9c03-28f866e02ed4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.479122] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1196.479122] env[62952]: value = "task-1263991" [ 1196.479122] env[62952]: _type = "Task" [ 1196.479122] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.485960] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263991, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.736946] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1196.736946] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ebc440-49a5-10eb-9377-e367ac2d6dea" [ 1196.736946] env[62952]: _type = "HttpNfcLease" [ 1196.736946] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1196.737295] env[62952]: DEBUG oslo_vmware.rw_handles [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1196.737295] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ebc440-49a5-10eb-9377-e367ac2d6dea" [ 1196.737295] env[62952]: _type = "HttpNfcLease" [ 1196.737295] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1196.738034] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00d5f41c-fd38-472a-b354-c7bd3df29789 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.745882] env[62952]: DEBUG oslo_vmware.rw_handles [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529dd2e6-4beb-872f-701e-9ec1cb3e90a7/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1196.746105] env[62952]: DEBUG oslo_vmware.rw_handles [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529dd2e6-4beb-872f-701e-9ec1cb3e90a7/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1196.809130] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263988, 'name': Rename_Task, 'duration_secs': 0.133278} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.809391] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1196.809627] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-61d8bd21-bf85-4b59-850f-713f8d833658 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.815066] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1196.815066] env[62952]: value = "task-1263992" [ 1196.815066] env[62952]: _type = "Task" [ 1196.815066] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.821810] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263992, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.830027] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6e5fddad-1cff-40d0-bc13-72142dffa1fb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.988980] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1263991, 'name': CreateVM_Task, 'duration_secs': 0.268975} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.989399] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1196.990247] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1196.990502] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1196.990913] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1196.991188] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c983e43-1ffb-4bd8-88b4-480f47a2796e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.995610] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1196.995610] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f117e2-8dfb-d81e-d722-c6832191ae0a" [ 1196.995610] env[62952]: _type = "Task" [ 1196.995610] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.003191] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f117e2-8dfb-d81e-d722-c6832191ae0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.324737] env[62952]: DEBUG oslo_vmware.api [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1263992, 'name': PowerOnVM_Task, 'duration_secs': 0.453138} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.325148] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1197.325353] env[62952]: INFO nova.compute.manager [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Took 7.66 seconds to spawn the instance on the hypervisor. [ 1197.325595] env[62952]: DEBUG nova.compute.manager [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1197.326481] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd05186-2ae0-4d61-9fd7-6b97490e1621 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.506722] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52f117e2-8dfb-d81e-d722-c6832191ae0a, 'name': SearchDatastore_Task, 'duration_secs': 0.009002} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.506722] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1197.507304] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1197.507712] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1197.507981] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1197.508701] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1197.508701] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bf34f694-a072-4bc6-968c-421361637037 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.516876] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1197.517247] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1197.518088] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8ec0120-a323-4708-a153-0ef82b69fea2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.523373] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1197.523373] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52367f3b-b8c7-e594-b04f-e0a05c6bd6b9" [ 1197.523373] env[62952]: _type = "Task" [ 1197.523373] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.531377] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52367f3b-b8c7-e594-b04f-e0a05c6bd6b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.845063] env[62952]: INFO nova.compute.manager [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Took 12.35 seconds to build instance. [ 1198.033906] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52367f3b-b8c7-e594-b04f-e0a05c6bd6b9, 'name': SearchDatastore_Task, 'duration_secs': 0.007698} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.034786] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-feee05f8-9fea-4fae-ba3d-0bd4a40dbdf1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.039853] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1198.039853] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e22b29-13a6-9a57-d868-82561f11bf73" [ 1198.039853] env[62952]: _type = "Task" [ 1198.039853] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.047215] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e22b29-13a6-9a57-d868-82561f11bf73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.347262] env[62952]: DEBUG oslo_concurrency.lockutils [None req-50c67f8a-8326-4e94-bc0c-141537026aa2 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.866s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.550735] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e22b29-13a6-9a57-d868-82561f11bf73, 'name': SearchDatastore_Task, 'duration_secs': 0.0099} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.551255] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1198.551625] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] c5d21282-1a4b-4998-8170-ee5423b958ad/c5d21282-1a4b-4998-8170-ee5423b958ad.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1198.552038] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0f58db42-f3d3-4ab2-af26-5a0df60d3878 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.559032] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1198.559032] env[62952]: value = "task-1263993" [ 1198.559032] env[62952]: _type = "Task" [ 1198.559032] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.566832] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1263993, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.991966] env[62952]: DEBUG nova.compute.manager [req-74176230-a1fe-4579-9d6f-a2904e8debcd req-1d2d964a-5bb9-4be5-96d0-5abf96f83839 service nova] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Received event network-changed-c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1198.992403] env[62952]: DEBUG nova.compute.manager [req-74176230-a1fe-4579-9d6f-a2904e8debcd req-1d2d964a-5bb9-4be5-96d0-5abf96f83839 service nova] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Refreshing instance network info cache due to event network-changed-c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1198.992827] env[62952]: DEBUG oslo_concurrency.lockutils [req-74176230-a1fe-4579-9d6f-a2904e8debcd req-1d2d964a-5bb9-4be5-96d0-5abf96f83839 service nova] Acquiring lock "refresh_cache-b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1198.993360] env[62952]: DEBUG oslo_concurrency.lockutils [req-74176230-a1fe-4579-9d6f-a2904e8debcd req-1d2d964a-5bb9-4be5-96d0-5abf96f83839 service nova] Acquired lock "refresh_cache-b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1198.993760] env[62952]: DEBUG nova.network.neutron [req-74176230-a1fe-4579-9d6f-a2904e8debcd req-1d2d964a-5bb9-4be5-96d0-5abf96f83839 service nova] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Refreshing network info cache for port c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1199.069752] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1263993, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457698} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.072841] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] c5d21282-1a4b-4998-8170-ee5423b958ad/c5d21282-1a4b-4998-8170-ee5423b958ad.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1199.072841] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1199.072841] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4572264e-58b4-45b9-b4cd-693d85358981 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.078032] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1199.078032] env[62952]: value = "task-1263994" [ 1199.078032] env[62952]: _type = "Task" [ 1199.078032] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.085666] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1263994, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.587198] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1263994, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067758} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.587535] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1199.588317] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b336839-7219-4063-b0ca-6926e55b02b0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.609613] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] c5d21282-1a4b-4998-8170-ee5423b958ad/c5d21282-1a4b-4998-8170-ee5423b958ad.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1199.612066] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2eb9d83-d8c1-4454-9f57-0992d60529f5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.631489] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1199.631489] env[62952]: value = "task-1263995" [ 1199.631489] env[62952]: _type = "Task" [ 1199.631489] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.639449] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1263995, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.784226] env[62952]: DEBUG nova.network.neutron [req-74176230-a1fe-4579-9d6f-a2904e8debcd req-1d2d964a-5bb9-4be5-96d0-5abf96f83839 service nova] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Updated VIF entry in instance network info cache for port c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1199.784646] env[62952]: DEBUG nova.network.neutron [req-74176230-a1fe-4579-9d6f-a2904e8debcd req-1d2d964a-5bb9-4be5-96d0-5abf96f83839 service nova] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Updating instance_info_cache with network_info: [{"id": "c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8", "address": "fa:16:3e:b0:e3:32", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3a25e8c-b6", "ovs_interfaceid": "c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1200.142116] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1263995, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.287295] env[62952]: DEBUG oslo_concurrency.lockutils [req-74176230-a1fe-4579-9d6f-a2904e8debcd req-1d2d964a-5bb9-4be5-96d0-5abf96f83839 service nova] Releasing lock "refresh_cache-b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1200.643173] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1263995, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.143305] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1263995, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.643926] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1263995, 'name': ReconfigVM_Task, 'duration_secs': 1.690719} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.644279] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Reconfigured VM instance instance-0000006b to attach disk [datastore1] c5d21282-1a4b-4998-8170-ee5423b958ad/c5d21282-1a4b-4998-8170-ee5423b958ad.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1201.644874] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-661d9068-767d-4c67-837a-adec7f64204c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.650855] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1201.650855] env[62952]: value = "task-1263996" [ 1201.650855] env[62952]: _type = "Task" [ 1201.650855] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.658174] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1263996, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.160780] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1263996, 'name': Rename_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.661164] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1263996, 'name': Rename_Task, 'duration_secs': 0.526674} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.661617] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1202.661698] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-215f1a35-99d4-43b4-b954-59a97b65002b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.667526] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1202.667526] env[62952]: value = "task-1263997" [ 1202.667526] env[62952]: _type = "Task" [ 1202.667526] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.676079] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1263997, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.176769] env[62952]: DEBUG oslo_vmware.api [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1263997, 'name': PowerOnVM_Task, 'duration_secs': 0.475949} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.177080] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1203.177294] env[62952]: INFO nova.compute.manager [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Took 9.78 seconds to spawn the instance on the hypervisor. [ 1203.177471] env[62952]: DEBUG nova.compute.manager [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1203.178217] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9ea5a4-30ae-4507-b2ef-b53a5988d078 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.327550] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.693916] env[62952]: INFO nova.compute.manager [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Took 14.47 seconds to build instance. [ 1203.831939] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Getting list of instances from cluster (obj){ [ 1203.831939] env[62952]: value = "domain-c8" [ 1203.831939] env[62952]: _type = "ClusterComputeResource" [ 1203.831939] env[62952]: } {{(pid=62952) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1203.833162] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98d27a53-e46d-479d-b24f-d7ddd60cfbfe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.844006] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Got total of 3 instances {{(pid=62952) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1203.844196] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Triggering sync for uuid f92ef782-fd64-47f0-8145-9261a1c3005c {{(pid=62952) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1203.844385] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Triggering sync for uuid b01ea5c5-1d49-47e7-ada2-acbacbe5abdb {{(pid=62952) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1203.844539] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Triggering sync for uuid c5d21282-1a4b-4998-8170-ee5423b958ad {{(pid=62952) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1203.844895] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "f92ef782-fd64-47f0-8145-9261a1c3005c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.845100] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "f92ef782-fd64-47f0-8145-9261a1c3005c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.845277] env[62952]: INFO nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] During sync_power_state the instance has a pending task (image_uploading). Skip. [ 1203.845442] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "f92ef782-fd64-47f0-8145-9261a1c3005c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.845635] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.845812] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.846110] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "c5d21282-1a4b-4998-8170-ee5423b958ad" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.846895] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9237e419-07ad-49d7-9073-e3e50ab4c343 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.196473] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a140bd2-c2eb-4177-9268-84f0c2ea73e2 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "c5d21282-1a4b-4998-8170-ee5423b958ad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.984s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1204.196779] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "c5d21282-1a4b-4998-8170-ee5423b958ad" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.351s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1204.197781] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a399274f-9d10-4d36-be9d-548088160904 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.355107] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.509s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1204.706897] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "c5d21282-1a4b-4998-8170-ee5423b958ad" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.510s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1204.734863] env[62952]: DEBUG nova.compute.manager [req-3535a96b-ebd9-44c1-853b-6ce2f3eb9b36 req-48d5ff51-a48a-4e22-b8be-2bf883bcaeae service nova] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Received event network-changed-f5eb1e66-2290-4179-b630-173dd8f3f058 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1204.734927] env[62952]: DEBUG nova.compute.manager [req-3535a96b-ebd9-44c1-853b-6ce2f3eb9b36 req-48d5ff51-a48a-4e22-b8be-2bf883bcaeae service nova] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Refreshing instance network info cache due to event network-changed-f5eb1e66-2290-4179-b630-173dd8f3f058. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1204.735155] env[62952]: DEBUG oslo_concurrency.lockutils [req-3535a96b-ebd9-44c1-853b-6ce2f3eb9b36 req-48d5ff51-a48a-4e22-b8be-2bf883bcaeae service nova] Acquiring lock "refresh_cache-c5d21282-1a4b-4998-8170-ee5423b958ad" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1204.735279] env[62952]: DEBUG oslo_concurrency.lockutils [req-3535a96b-ebd9-44c1-853b-6ce2f3eb9b36 req-48d5ff51-a48a-4e22-b8be-2bf883bcaeae service nova] Acquired lock "refresh_cache-c5d21282-1a4b-4998-8170-ee5423b958ad" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1204.735428] env[62952]: DEBUG nova.network.neutron [req-3535a96b-ebd9-44c1-853b-6ce2f3eb9b36 req-48d5ff51-a48a-4e22-b8be-2bf883bcaeae service nova] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Refreshing network info cache for port f5eb1e66-2290-4179-b630-173dd8f3f058 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1205.454843] env[62952]: DEBUG nova.network.neutron [req-3535a96b-ebd9-44c1-853b-6ce2f3eb9b36 req-48d5ff51-a48a-4e22-b8be-2bf883bcaeae service nova] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Updated VIF entry in instance network info cache for port f5eb1e66-2290-4179-b630-173dd8f3f058. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1205.455241] env[62952]: DEBUG nova.network.neutron [req-3535a96b-ebd9-44c1-853b-6ce2f3eb9b36 req-48d5ff51-a48a-4e22-b8be-2bf883bcaeae service nova] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Updating instance_info_cache with network_info: [{"id": "f5eb1e66-2290-4179-b630-173dd8f3f058", "address": "fa:16:3e:15:7b:1a", "network": {"id": "8e167a8b-2628-43e1-8418-e673064745b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-232555460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f725be1feb8a44478d297b0be58c59ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5eb1e66-22", "ovs_interfaceid": "f5eb1e66-2290-4179-b630-173dd8f3f058", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1205.958861] env[62952]: DEBUG oslo_concurrency.lockutils [req-3535a96b-ebd9-44c1-853b-6ce2f3eb9b36 req-48d5ff51-a48a-4e22-b8be-2bf883bcaeae service nova] Releasing lock "refresh_cache-c5d21282-1a4b-4998-8170-ee5423b958ad" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1206.272660] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1206.776132] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1206.776394] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1206.776604] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1206.776831] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1206.777790] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91586275-68c9-4ac7-9af9-b19fd72e77bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.785914] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8cf9f3-e1bc-44be-809b-0c27a145a609 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.801515] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843eca5d-d731-42e1-bbc3-2332e45f0f1e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.807871] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c58124-27ce-4d5e-b3b3-1fc970f95d63 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.836082] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180832MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1206.836220] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1206.836418] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.862637] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance f92ef782-fd64-47f0-8145-9261a1c3005c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1207.862962] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance b01ea5c5-1d49-47e7-ada2-acbacbe5abdb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1207.862962] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance c5d21282-1a4b-4998-8170-ee5423b958ad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1207.863088] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1207.863229] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1207.914575] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0275a72-b570-4464-a7f8-b215ec75eadd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.922439] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e07541b9-035b-499c-82fa-3ed98f6f5c56 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.953275] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-699fd73e-f676-462f-9263-b5edd2f7c404 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.960881] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b2459d-e4c0-46f9-8acf-30b29deee52b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.974378] env[62952]: DEBUG nova.compute.provider_tree [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1208.155708] env[62952]: DEBUG oslo_vmware.rw_handles [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529dd2e6-4beb-872f-701e-9ec1cb3e90a7/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1208.156705] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcf5ebca-df1d-450e-ae55-7a84e1c8f3ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.163123] env[62952]: DEBUG oslo_vmware.rw_handles [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529dd2e6-4beb-872f-701e-9ec1cb3e90a7/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1208.163295] env[62952]: ERROR oslo_vmware.rw_handles [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529dd2e6-4beb-872f-701e-9ec1cb3e90a7/disk-0.vmdk due to incomplete transfer. [ 1208.163507] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-de62a038-f62f-4dff-83d6-62da423efbda {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.170133] env[62952]: DEBUG oslo_vmware.rw_handles [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529dd2e6-4beb-872f-701e-9ec1cb3e90a7/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1208.170338] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Uploaded image 4bad21a1-8a67-4b7d-acb8-57b395177cc7 to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1208.172564] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1208.172797] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-34d9e8e1-3f8b-402f-8796-7df76ea52e74 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.178082] env[62952]: DEBUG oslo_vmware.api [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1208.178082] env[62952]: value = "task-1263998" [ 1208.178082] env[62952]: _type = "Task" [ 1208.178082] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.185477] env[62952]: DEBUG oslo_vmware.api [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263998, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.477288] env[62952]: DEBUG nova.scheduler.client.report [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1208.690131] env[62952]: DEBUG oslo_vmware.api [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263998, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.982058] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1208.982416] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.146s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.189384] env[62952]: DEBUG oslo_vmware.api [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263998, 'name': Destroy_Task} progress is 33%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.689673] env[62952]: DEBUG oslo_vmware.api [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263998, 'name': Destroy_Task, 'duration_secs': 1.19852} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.690044] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Destroyed the VM [ 1209.690335] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1209.690620] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-cf835893-1fee-4eaa-9acc-b6077b074cf2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.697154] env[62952]: DEBUG oslo_vmware.api [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1209.697154] env[62952]: value = "task-1263999" [ 1209.697154] env[62952]: _type = "Task" [ 1209.697154] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.704063] env[62952]: DEBUG oslo_vmware.api [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263999, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.207301] env[62952]: DEBUG oslo_vmware.api [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1263999, 'name': RemoveSnapshot_Task, 'duration_secs': 0.433356} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.207667] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1210.207833] env[62952]: INFO nova.compute.manager [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Took 17.82 seconds to snapshot the instance on the hypervisor. [ 1210.756013] env[62952]: DEBUG nova.compute.manager [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Found 3 images (rotation: 2) {{(pid=62952) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1210.756231] env[62952]: DEBUG nova.compute.manager [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Rotating out 1 backups {{(pid=62952) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 1210.756397] env[62952]: DEBUG nova.compute.manager [None req-68cc5c08-85ef-4f39-b194-b6b9e74c45a0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Deleting image efbe8091-a483-4979-a9f8-af2c3c2a058c {{(pid=62952) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 1210.981910] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1210.982164] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1210.982164] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1211.541565] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1211.541565] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquired lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1211.541565] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Forcefully refreshing network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1211.541565] env[62952]: DEBUG nova.objects.instance [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lazy-loading 'info_cache' on Instance uuid f92ef782-fd64-47f0-8145-9261a1c3005c {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1213.260982] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aba69d38-b3d7-41ae-bd8d-a669377f9886 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "f92ef782-fd64-47f0-8145-9261a1c3005c" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1213.261369] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aba69d38-b3d7-41ae-bd8d-a669377f9886 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "f92ef782-fd64-47f0-8145-9261a1c3005c" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.261449] env[62952]: DEBUG nova.compute.manager [None req-aba69d38-b3d7-41ae-bd8d-a669377f9886 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1213.262351] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbba203f-1cd4-4037-b90d-37c3ae5a42b9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.269435] env[62952]: DEBUG nova.compute.manager [None req-aba69d38-b3d7-41ae-bd8d-a669377f9886 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62952) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1213.269987] env[62952]: DEBUG nova.objects.instance [None req-aba69d38-b3d7-41ae-bd8d-a669377f9886 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lazy-loading 'flavor' on Instance uuid f92ef782-fd64-47f0-8145-9261a1c3005c {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1213.281454] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance_info_cache with network_info: [{"id": "19fcce34-055a-492e-8c73-a39767f93814", "address": "fa:16:3e:71:3d:5f", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19fcce34-05", "ovs_interfaceid": "19fcce34-055a-492e-8c73-a39767f93814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1213.775086] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-aba69d38-b3d7-41ae-bd8d-a669377f9886 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1213.775375] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f951bd76-f199-4cff-a52a-a05620ad7983 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.783442] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Releasing lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1213.783622] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updated the network info_cache for instance {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1213.783901] env[62952]: DEBUG oslo_vmware.api [None req-aba69d38-b3d7-41ae-bd8d-a669377f9886 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1213.783901] env[62952]: value = "task-1264000" [ 1213.783901] env[62952]: _type = "Task" [ 1213.783901] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.784116] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1213.784325] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1213.787265] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1213.787438] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1213.787957] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1213.788141] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1213.788290] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1213.793027] env[62952]: DEBUG oslo_vmware.api [None req-aba69d38-b3d7-41ae-bd8d-a669377f9886 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264000, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.293451] env[62952]: DEBUG oslo_vmware.api [None req-aba69d38-b3d7-41ae-bd8d-a669377f9886 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264000, 'name': PowerOffVM_Task, 'duration_secs': 0.234514} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.293912] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-aba69d38-b3d7-41ae-bd8d-a669377f9886 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1214.293912] env[62952]: DEBUG nova.compute.manager [None req-aba69d38-b3d7-41ae-bd8d-a669377f9886 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1214.294657] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fd2feb9-9474-46d4-81a7-866e5d2ffabd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.805900] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aba69d38-b3d7-41ae-bd8d-a669377f9886 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "f92ef782-fd64-47f0-8145-9261a1c3005c" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.544s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1216.370801] env[62952]: DEBUG nova.compute.manager [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Stashing vm_state: stopped {{(pid=62952) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1216.890154] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1216.890439] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1217.074763] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1217.396730] env[62952]: INFO nova.compute.claims [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1217.903434] env[62952]: INFO nova.compute.resource_tracker [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating resource usage from migration 83bcd19f-f31e-4099-a5da-1ea7553f8868 [ 1217.972662] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d68b71c-56a1-463f-af03-a200883e0e27 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.980305] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d07d368-9b3c-4936-99a6-62dad1c9c0ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.018022] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b0273f-c90a-4603-8a23-de453a174189 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.027484] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f248099-997f-4652-a349-7a9d1e920bb2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.045936] env[62952]: DEBUG nova.compute.provider_tree [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1218.549256] env[62952]: DEBUG nova.scheduler.client.report [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1219.055236] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.165s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1219.055486] env[62952]: INFO nova.compute.manager [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Migrating [ 1219.570309] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1219.570681] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1219.570681] env[62952]: DEBUG nova.network.neutron [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1220.289185] env[62952]: DEBUG nova.network.neutron [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance_info_cache with network_info: [{"id": "19fcce34-055a-492e-8c73-a39767f93814", "address": "fa:16:3e:71:3d:5f", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19fcce34-05", "ovs_interfaceid": "19fcce34-055a-492e-8c73-a39767f93814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1220.791558] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1222.307364] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6629a7ad-690b-47e0-a634-60b0c23d26e3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.325654] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance 'f92ef782-fd64-47f0-8145-9261a1c3005c' progress to 0 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1222.831751] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1222.832080] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f0f4c0bf-85f2-4aa2-8405-6e5956919f18 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.839839] env[62952]: DEBUG oslo_vmware.api [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1222.839839] env[62952]: value = "task-1264001" [ 1222.839839] env[62952]: _type = "Task" [ 1222.839839] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.847162] env[62952]: DEBUG oslo_vmware.api [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264001, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.351115] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1223.351499] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance 'f92ef782-fd64-47f0-8145-9261a1c3005c' progress to 17 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1223.857572] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1223.857814] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1223.857943] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1223.858153] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1223.858303] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1223.858455] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1223.858661] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1223.858822] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1223.858987] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1223.859170] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1223.859348] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1223.864245] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df285cac-7cce-476b-adfd-fcbea4ddfa25 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.880560] env[62952]: DEBUG oslo_vmware.api [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1223.880560] env[62952]: value = "task-1264002" [ 1223.880560] env[62952]: _type = "Task" [ 1223.880560] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.888455] env[62952]: DEBUG oslo_vmware.api [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264002, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.390959] env[62952]: DEBUG oslo_vmware.api [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264002, 'name': ReconfigVM_Task, 'duration_secs': 0.13623} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.391303] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance 'f92ef782-fd64-47f0-8145-9261a1c3005c' progress to 33 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1224.898106] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1224.898446] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1224.898560] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1224.898751] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1224.898908] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1224.899159] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1224.899272] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1224.899439] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1224.899632] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1224.899756] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1224.899930] env[62952]: DEBUG nova.virt.hardware [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1224.905135] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Reconfiguring VM instance instance-00000068 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1224.905424] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50b622ab-9f5f-4407-85c5-310f97a97d61 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.923998] env[62952]: DEBUG oslo_vmware.api [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1224.923998] env[62952]: value = "task-1264003" [ 1224.923998] env[62952]: _type = "Task" [ 1224.923998] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.931520] env[62952]: DEBUG oslo_vmware.api [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264003, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.434204] env[62952]: DEBUG oslo_vmware.api [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264003, 'name': ReconfigVM_Task, 'duration_secs': 0.150448} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.434545] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Reconfigured VM instance instance-00000068 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1225.435217] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6695952c-cf6f-4b6b-be9e-74f9d98943b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.456818] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] f92ef782-fd64-47f0-8145-9261a1c3005c/f92ef782-fd64-47f0-8145-9261a1c3005c.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1225.457432] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b7665bf-fbce-4772-a567-41b6742e17f1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.475238] env[62952]: DEBUG oslo_vmware.api [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1225.475238] env[62952]: value = "task-1264004" [ 1225.475238] env[62952]: _type = "Task" [ 1225.475238] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.483758] env[62952]: DEBUG oslo_vmware.api [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264004, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.985072] env[62952]: DEBUG oslo_vmware.api [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264004, 'name': ReconfigVM_Task, 'duration_secs': 0.242175} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.985377] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Reconfigured VM instance instance-00000068 to attach disk [datastore2] f92ef782-fd64-47f0-8145-9261a1c3005c/f92ef782-fd64-47f0-8145-9261a1c3005c.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1225.985653] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance 'f92ef782-fd64-47f0-8145-9261a1c3005c' progress to 50 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1226.492272] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47fcd0e9-1212-4dcb-8cb4-e1153f865438 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.511855] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b94e694-c222-4c3c-99ad-eabd03c821a1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.528331] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance 'f92ef782-fd64-47f0-8145-9261a1c3005c' progress to 67 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1227.069980] env[62952]: DEBUG nova.network.neutron [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Port 19fcce34-055a-492e-8c73-a39767f93814 binding to destination host cpu-1 is already ACTIVE {{(pid=62952) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1228.094952] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "f92ef782-fd64-47f0-8145-9261a1c3005c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1228.095272] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "f92ef782-fd64-47f0-8145-9261a1c3005c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1228.095412] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "f92ef782-fd64-47f0-8145-9261a1c3005c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.141705] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1229.141940] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1229.142096] env[62952]: DEBUG nova.network.neutron [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1229.893074] env[62952]: DEBUG nova.network.neutron [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance_info_cache with network_info: [{"id": "19fcce34-055a-492e-8c73-a39767f93814", "address": "fa:16:3e:71:3d:5f", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19fcce34-05", "ovs_interfaceid": "19fcce34-055a-492e-8c73-a39767f93814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1230.396569] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1230.922368] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5701750c-abb2-4870-ab10-645ab69ea25a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.941050] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce9bb0d9-45cc-43cf-9e38-d8e87ce69b69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.947894] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance 'f92ef782-fd64-47f0-8145-9261a1c3005c' progress to 83 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1231.453917] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eb80f41b-0481-4dbb-895d-b21a9659af7e tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance 'f92ef782-fd64-47f0-8145-9261a1c3005c' progress to 100 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1233.633115] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "f92ef782-fd64-47f0-8145-9261a1c3005c" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1233.633465] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "f92ef782-fd64-47f0-8145-9261a1c3005c" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1233.633597] env[62952]: DEBUG nova.compute.manager [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Going to confirm migration 6 {{(pid=62952) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1234.177564] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1234.177756] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1234.177937] env[62952]: DEBUG nova.network.neutron [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1234.178197] env[62952]: DEBUG nova.objects.instance [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lazy-loading 'info_cache' on Instance uuid f92ef782-fd64-47f0-8145-9261a1c3005c {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1235.090493] env[62952]: DEBUG oslo_concurrency.lockutils [None req-34789e00-c539-4a3f-b8c2-5595bca160b9 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1235.090879] env[62952]: DEBUG oslo_concurrency.lockutils [None req-34789e00-c539-4a3f-b8c2-5595bca160b9 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1235.090986] env[62952]: DEBUG nova.compute.manager [None req-34789e00-c539-4a3f-b8c2-5595bca160b9 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1235.091917] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e1dd971-0948-48aa-ad70-662f5b63bb4b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.099097] env[62952]: DEBUG nova.compute.manager [None req-34789e00-c539-4a3f-b8c2-5595bca160b9 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62952) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1235.099669] env[62952]: DEBUG nova.objects.instance [None req-34789e00-c539-4a3f-b8c2-5595bca160b9 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lazy-loading 'flavor' on Instance uuid b01ea5c5-1d49-47e7-ada2-acbacbe5abdb {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1235.449811] env[62952]: DEBUG nova.network.neutron [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance_info_cache with network_info: [{"id": "19fcce34-055a-492e-8c73-a39767f93814", "address": "fa:16:3e:71:3d:5f", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19fcce34-05", "ovs_interfaceid": "19fcce34-055a-492e-8c73-a39767f93814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1235.604748] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-34789e00-c539-4a3f-b8c2-5595bca160b9 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1235.604980] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a693195-de97-4e0f-95a5-b134cc0a5d0c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.611600] env[62952]: DEBUG oslo_vmware.api [None req-34789e00-c539-4a3f-b8c2-5595bca160b9 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1235.611600] env[62952]: value = "task-1264005" [ 1235.611600] env[62952]: _type = "Task" [ 1235.611600] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.619411] env[62952]: DEBUG oslo_vmware.api [None req-34789e00-c539-4a3f-b8c2-5595bca160b9 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1264005, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.952493] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1235.952791] env[62952]: DEBUG nova.objects.instance [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lazy-loading 'migration_context' on Instance uuid f92ef782-fd64-47f0-8145-9261a1c3005c {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1236.121074] env[62952]: DEBUG oslo_vmware.api [None req-34789e00-c539-4a3f-b8c2-5595bca160b9 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1264005, 'name': PowerOffVM_Task, 'duration_secs': 0.150044} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.121456] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-34789e00-c539-4a3f-b8c2-5595bca160b9 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1236.121623] env[62952]: DEBUG nova.compute.manager [None req-34789e00-c539-4a3f-b8c2-5595bca160b9 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1236.122394] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ebaf20-1098-4040-8a19-918ad6f79e05 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.456417] env[62952]: DEBUG nova.objects.base [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1236.457341] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ce2c3f-7a7d-4666-a9c8-9e50fca23dc1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.476341] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-869f4453-0c8d-4b64-bec1-eb85d4132c9a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.481259] env[62952]: DEBUG oslo_vmware.api [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1236.481259] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]528aded3-f2cb-c51d-8d8c-c0d403d168b9" [ 1236.481259] env[62952]: _type = "Task" [ 1236.481259] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.488442] env[62952]: DEBUG oslo_vmware.api [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]528aded3-f2cb-c51d-8d8c-c0d403d168b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.633813] env[62952]: DEBUG oslo_concurrency.lockutils [None req-34789e00-c539-4a3f-b8c2-5595bca160b9 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.543s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1236.992050] env[62952]: DEBUG oslo_vmware.api [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]528aded3-f2cb-c51d-8d8c-c0d403d168b9, 'name': SearchDatastore_Task, 'duration_secs': 0.009349} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.992050] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1236.992050] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1237.466387] env[62952]: DEBUG nova.objects.instance [None req-ab142a05-a4ae-456a-99b6-4ee8d199d6c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lazy-loading 'flavor' on Instance uuid b01ea5c5-1d49-47e7-ada2-acbacbe5abdb {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1237.554401] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29a21e5-37a7-4660-a213-b25c9569a76b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.561741] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba47bbd-374a-42df-9cd7-feba25ba231c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.590491] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44f8280-f21b-400d-94f8-6c5b583e4aca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.597270] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98d43f48-da99-486d-9696-2ea1869218ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.609825] env[62952]: DEBUG nova.compute.provider_tree [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1237.971326] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab142a05-a4ae-456a-99b6-4ee8d199d6c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "refresh_cache-b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1237.971506] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab142a05-a4ae-456a-99b6-4ee8d199d6c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "refresh_cache-b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1237.971667] env[62952]: DEBUG nova.network.neutron [None req-ab142a05-a4ae-456a-99b6-4ee8d199d6c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1237.971849] env[62952]: DEBUG nova.objects.instance [None req-ab142a05-a4ae-456a-99b6-4ee8d199d6c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lazy-loading 'info_cache' on Instance uuid b01ea5c5-1d49-47e7-ada2-acbacbe5abdb {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1238.113610] env[62952]: DEBUG nova.scheduler.client.report [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1238.475823] env[62952]: DEBUG nova.objects.base [None req-ab142a05-a4ae-456a-99b6-4ee8d199d6c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1239.123118] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.131s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1239.123347] env[62952]: DEBUG nova.compute.manager [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=62952) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 1239.173147] env[62952]: DEBUG nova.network.neutron [None req-ab142a05-a4ae-456a-99b6-4ee8d199d6c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Updating instance_info_cache with network_info: [{"id": "c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8", "address": "fa:16:3e:b0:e3:32", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3a25e8c-b6", "ovs_interfaceid": "c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1239.676132] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab142a05-a4ae-456a-99b6-4ee8d199d6c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "refresh_cache-b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1239.678942] env[62952]: INFO nova.scheduler.client.report [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Deleted allocation for migration 83bcd19f-f31e-4099-a5da-1ea7553f8868 [ 1240.181145] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab142a05-a4ae-456a-99b6-4ee8d199d6c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1240.181453] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e932caa-d159-4bee-9063-3db528d36ed9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.184255] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4d3f17a6-d672-40d5-9a27-df19df6a4b94 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "f92ef782-fd64-47f0-8145-9261a1c3005c" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.551s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.190070] env[62952]: DEBUG oslo_vmware.api [None req-ab142a05-a4ae-456a-99b6-4ee8d199d6c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1240.190070] env[62952]: value = "task-1264006" [ 1240.190070] env[62952]: _type = "Task" [ 1240.190070] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.197271] env[62952]: DEBUG oslo_vmware.api [None req-ab142a05-a4ae-456a-99b6-4ee8d199d6c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1264006, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.589932] env[62952]: DEBUG nova.objects.instance [None req-5b6cc46f-7932-430a-af35-7b246ea81ddd tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lazy-loading 'flavor' on Instance uuid f92ef782-fd64-47f0-8145-9261a1c3005c {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1240.698635] env[62952]: DEBUG oslo_vmware.api [None req-ab142a05-a4ae-456a-99b6-4ee8d199d6c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1264006, 'name': PowerOnVM_Task, 'duration_secs': 0.36968} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.699073] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab142a05-a4ae-456a-99b6-4ee8d199d6c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1240.699073] env[62952]: DEBUG nova.compute.manager [None req-ab142a05-a4ae-456a-99b6-4ee8d199d6c0 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1240.699779] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9acb9b49-225d-4c4a-b802-4d332b0df932 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.094796] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b6cc46f-7932-430a-af35-7b246ea81ddd tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1241.094984] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b6cc46f-7932-430a-af35-7b246ea81ddd tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1241.095179] env[62952]: DEBUG nova.network.neutron [None req-5b6cc46f-7932-430a-af35-7b246ea81ddd tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1241.095363] env[62952]: DEBUG nova.objects.instance [None req-5b6cc46f-7932-430a-af35-7b246ea81ddd tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lazy-loading 'info_cache' on Instance uuid f92ef782-fd64-47f0-8145-9261a1c3005c {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1241.598502] env[62952]: DEBUG nova.objects.base [None req-5b6cc46f-7932-430a-af35-7b246ea81ddd tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1242.337835] env[62952]: DEBUG nova.network.neutron [None req-5b6cc46f-7932-430a-af35-7b246ea81ddd tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance_info_cache with network_info: [{"id": "19fcce34-055a-492e-8c73-a39767f93814", "address": "fa:16:3e:71:3d:5f", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19fcce34-05", "ovs_interfaceid": "19fcce34-055a-492e-8c73-a39767f93814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.404449] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "c5d21282-1a4b-4998-8170-ee5423b958ad" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1242.404707] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "c5d21282-1a4b-4998-8170-ee5423b958ad" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1242.444152] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce61f832-54e6-47ca-b004-eb0834db5e6b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.450897] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-320a8b67-5862-473b-86cb-20af9621b275 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Suspending the VM {{(pid=62952) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1242.451175] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-9fac014f-37cb-462c-8c68-e93f447ec524 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.456980] env[62952]: DEBUG oslo_vmware.api [None req-320a8b67-5862-473b-86cb-20af9621b275 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1242.456980] env[62952]: value = "task-1264007" [ 1242.456980] env[62952]: _type = "Task" [ 1242.456980] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.464544] env[62952]: DEBUG oslo_vmware.api [None req-320a8b67-5862-473b-86cb-20af9621b275 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1264007, 'name': SuspendVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.840526] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b6cc46f-7932-430a-af35-7b246ea81ddd tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "refresh_cache-f92ef782-fd64-47f0-8145-9261a1c3005c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1242.908018] env[62952]: DEBUG nova.compute.utils [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1242.966609] env[62952]: DEBUG oslo_vmware.api [None req-320a8b67-5862-473b-86cb-20af9621b275 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1264007, 'name': SuspendVM_Task} progress is 70%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.343643] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b6cc46f-7932-430a-af35-7b246ea81ddd tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1243.344065] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32c7faa2-d2d1-4797-9bbc-4c4b5f91e9f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.351702] env[62952]: DEBUG oslo_vmware.api [None req-5b6cc46f-7932-430a-af35-7b246ea81ddd tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1243.351702] env[62952]: value = "task-1264008" [ 1243.351702] env[62952]: _type = "Task" [ 1243.351702] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.359569] env[62952]: DEBUG oslo_vmware.api [None req-5b6cc46f-7932-430a-af35-7b246ea81ddd tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264008, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.410520] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "c5d21282-1a4b-4998-8170-ee5423b958ad" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1243.466761] env[62952]: DEBUG oslo_vmware.api [None req-320a8b67-5862-473b-86cb-20af9621b275 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1264007, 'name': SuspendVM_Task, 'duration_secs': 0.615668} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.467032] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-320a8b67-5862-473b-86cb-20af9621b275 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Suspended the VM {{(pid=62952) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1243.467217] env[62952]: DEBUG nova.compute.manager [None req-320a8b67-5862-473b-86cb-20af9621b275 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1243.467972] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d457d7-fdb6-475f-99a6-cdf59cc6c698 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.861845] env[62952]: DEBUG oslo_vmware.api [None req-5b6cc46f-7932-430a-af35-7b246ea81ddd tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264008, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.362284] env[62952]: DEBUG oslo_vmware.api [None req-5b6cc46f-7932-430a-af35-7b246ea81ddd tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264008, 'name': PowerOnVM_Task, 'duration_secs': 0.791256} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.362611] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b6cc46f-7932-430a-af35-7b246ea81ddd tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1244.362781] env[62952]: DEBUG nova.compute.manager [None req-5b6cc46f-7932-430a-af35-7b246ea81ddd tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1244.363579] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed435753-cc86-4022-91f4-329b8347c47e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.469754] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "c5d21282-1a4b-4998-8170-ee5423b958ad" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1244.470020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "c5d21282-1a4b-4998-8170-ee5423b958ad" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1244.470261] env[62952]: INFO nova.compute.manager [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Attaching volume 69b2eba4-89f3-4502-ac1e-cac959b2e708 to /dev/sdb [ 1244.499556] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb75b20f-797d-4b0b-9216-d134bae7e035 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.506462] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ad983d7-525a-414b-ae84-d16f99ebfd36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.518941] env[62952]: DEBUG nova.virt.block_device [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Updating existing volume attachment record: 9d59cf89-eb31-4018-a690-5f9735473dba {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1244.797490] env[62952]: INFO nova.compute.manager [None req-e5aca491-2bf6-4e83-8752-6f78c8152b13 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Resuming [ 1244.798178] env[62952]: DEBUG nova.objects.instance [None req-e5aca491-2bf6-4e83-8752-6f78c8152b13 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lazy-loading 'flavor' on Instance uuid b01ea5c5-1d49-47e7-ada2-acbacbe5abdb {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1245.185782] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "f92ef782-fd64-47f0-8145-9261a1c3005c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1245.186099] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "f92ef782-fd64-47f0-8145-9261a1c3005c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1245.186332] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "f92ef782-fd64-47f0-8145-9261a1c3005c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1245.186525] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "f92ef782-fd64-47f0-8145-9261a1c3005c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1245.186700] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "f92ef782-fd64-47f0-8145-9261a1c3005c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1245.188895] env[62952]: INFO nova.compute.manager [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Terminating instance [ 1245.190606] env[62952]: DEBUG nova.compute.manager [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1245.190808] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1245.191646] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2185560f-1efe-428b-9a20-59777362cb0b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.200885] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1245.201169] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-61e76d84-c87f-4e88-81e5-0e3172b83bac {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.207460] env[62952]: DEBUG oslo_vmware.api [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1245.207460] env[62952]: value = "task-1264012" [ 1245.207460] env[62952]: _type = "Task" [ 1245.207460] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.214581] env[62952]: DEBUG oslo_vmware.api [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264012, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.717909] env[62952]: DEBUG oslo_vmware.api [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264012, 'name': PowerOffVM_Task, 'duration_secs': 0.172818} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.718294] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1245.718355] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1245.718600] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-20fa2a97-6940-4fea-8fb1-5b48c357a62f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.775725] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1245.775947] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1245.776180] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Deleting the datastore file [datastore2] f92ef782-fd64-47f0-8145-9261a1c3005c {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1245.776466] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8d39029c-07a7-487b-9138-375df4355a5a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.782699] env[62952]: DEBUG oslo_vmware.api [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1245.782699] env[62952]: value = "task-1264014" [ 1245.782699] env[62952]: _type = "Task" [ 1245.782699] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.790268] env[62952]: DEBUG oslo_vmware.api [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264014, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.809417] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5aca491-2bf6-4e83-8752-6f78c8152b13 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "refresh_cache-b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1245.809599] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5aca491-2bf6-4e83-8752-6f78c8152b13 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquired lock "refresh_cache-b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1245.809775] env[62952]: DEBUG nova.network.neutron [None req-e5aca491-2bf6-4e83-8752-6f78c8152b13 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1246.292635] env[62952]: DEBUG oslo_vmware.api [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264014, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142251} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.292852] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1246.293054] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1246.293244] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1246.293424] env[62952]: INFO nova.compute.manager [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1246.293665] env[62952]: DEBUG oslo.service.loopingcall [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1246.293862] env[62952]: DEBUG nova.compute.manager [-] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1246.293958] env[62952]: DEBUG nova.network.neutron [-] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1246.512319] env[62952]: DEBUG nova.network.neutron [None req-e5aca491-2bf6-4e83-8752-6f78c8152b13 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Updating instance_info_cache with network_info: [{"id": "c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8", "address": "fa:16:3e:b0:e3:32", "network": {"id": "e5d1b7dd-b2d0-4153-9a49-68c347148d33", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-993379918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5080fbeac0744478843ab28efea8fc18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3a25e8c-b6", "ovs_interfaceid": "c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1246.747339] env[62952]: DEBUG nova.compute.manager [req-18b67a3c-8be4-4e61-80a7-313bc2fbf3b5 req-2ef693ca-133d-4475-aa65-5a961c61a68c service nova] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Received event network-vif-deleted-19fcce34-055a-492e-8c73-a39767f93814 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1246.747594] env[62952]: INFO nova.compute.manager [req-18b67a3c-8be4-4e61-80a7-313bc2fbf3b5 req-2ef693ca-133d-4475-aa65-5a961c61a68c service nova] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Neutron deleted interface 19fcce34-055a-492e-8c73-a39767f93814; detaching it from the instance and deleting it from the info cache [ 1246.747695] env[62952]: DEBUG nova.network.neutron [req-18b67a3c-8be4-4e61-80a7-313bc2fbf3b5 req-2ef693ca-133d-4475-aa65-5a961c61a68c service nova] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1247.014939] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5aca491-2bf6-4e83-8752-6f78c8152b13 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Releasing lock "refresh_cache-b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1247.016355] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3a001c6-7eaf-406b-9452-8dd0fba73743 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.022780] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e5aca491-2bf6-4e83-8752-6f78c8152b13 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Resuming the VM {{(pid=62952) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1247.023115] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef99e819-0509-4c54-8eea-727e997036f3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.028944] env[62952]: DEBUG oslo_vmware.api [None req-e5aca491-2bf6-4e83-8752-6f78c8152b13 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1247.028944] env[62952]: value = "task-1264016" [ 1247.028944] env[62952]: _type = "Task" [ 1247.028944] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.035914] env[62952]: DEBUG oslo_vmware.api [None req-e5aca491-2bf6-4e83-8752-6f78c8152b13 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1264016, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.228026] env[62952]: DEBUG nova.network.neutron [-] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1247.250646] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9201e0d5-f5fe-4a1f-b966-2b5be0eec2d4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.259934] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e94c66-3ff2-488d-90e3-82e50551208b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.283626] env[62952]: DEBUG nova.compute.manager [req-18b67a3c-8be4-4e61-80a7-313bc2fbf3b5 req-2ef693ca-133d-4475-aa65-5a961c61a68c service nova] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Detach interface failed, port_id=19fcce34-055a-492e-8c73-a39767f93814, reason: Instance f92ef782-fd64-47f0-8145-9261a1c3005c could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1247.539881] env[62952]: DEBUG oslo_vmware.api [None req-e5aca491-2bf6-4e83-8752-6f78c8152b13 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1264016, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.730907] env[62952]: INFO nova.compute.manager [-] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Took 1.44 seconds to deallocate network for instance. [ 1248.038994] env[62952]: DEBUG oslo_vmware.api [None req-e5aca491-2bf6-4e83-8752-6f78c8152b13 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1264016, 'name': PowerOnVM_Task, 'duration_secs': 0.545693} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.039301] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e5aca491-2bf6-4e83-8752-6f78c8152b13 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Resumed the VM {{(pid=62952) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1248.039475] env[62952]: DEBUG nova.compute.manager [None req-e5aca491-2bf6-4e83-8752-6f78c8152b13 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1248.040270] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e4b68fd-b059-47a4-813c-d805e36f2fbe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.237857] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1248.238142] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1248.238350] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1248.265721] env[62952]: INFO nova.scheduler.client.report [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Deleted allocations for instance f92ef782-fd64-47f0-8145-9261a1c3005c [ 1248.773856] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3f2c1493-e134-4d15-9326-0f998f613ea0 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "f92ef782-fd64-47f0-8145-9261a1c3005c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.588s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.061679] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1249.061920] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271995', 'volume_id': '69b2eba4-89f3-4502-ac1e-cac959b2e708', 'name': 'volume-69b2eba4-89f3-4502-ac1e-cac959b2e708', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c5d21282-1a4b-4998-8170-ee5423b958ad', 'attached_at': '', 'detached_at': '', 'volume_id': '69b2eba4-89f3-4502-ac1e-cac959b2e708', 'serial': '69b2eba4-89f3-4502-ac1e-cac959b2e708'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1249.062830] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1860c5dc-3098-4f66-84a5-7844354bdf48 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.080815] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5912dd7f-0782-493f-8280-e6ea1df5e6f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.105702] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] volume-69b2eba4-89f3-4502-ac1e-cac959b2e708/volume-69b2eba4-89f3-4502-ac1e-cac959b2e708.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1249.106046] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0596da5-a15d-4630-bc28-5d703dbb8a09 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.124908] env[62952]: DEBUG oslo_vmware.api [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1249.124908] env[62952]: value = "task-1264017" [ 1249.124908] env[62952]: _type = "Task" [ 1249.124908] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.134646] env[62952]: DEBUG oslo_vmware.api [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264017, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.561896] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.562230] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.562611] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.562856] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.563060] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.565443] env[62952]: INFO nova.compute.manager [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Terminating instance [ 1249.567393] env[62952]: DEBUG nova.compute.manager [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1249.567594] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1249.568474] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a65491-a787-4a6c-b650-754f602e82d9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.576131] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1249.576380] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d29d2652-34de-435e-bdd1-477a455b6a30 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.582772] env[62952]: DEBUG oslo_vmware.api [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1249.582772] env[62952]: value = "task-1264018" [ 1249.582772] env[62952]: _type = "Task" [ 1249.582772] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.590238] env[62952]: DEBUG oslo_vmware.api [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1264018, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.634566] env[62952]: DEBUG oslo_vmware.api [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264017, 'name': ReconfigVM_Task, 'duration_secs': 0.491798} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.634835] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Reconfigured VM instance instance-0000006b to attach disk [datastore2] volume-69b2eba4-89f3-4502-ac1e-cac959b2e708/volume-69b2eba4-89f3-4502-ac1e-cac959b2e708.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1249.639569] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88bbfe4f-91c2-4e07-ade6-948c30d76a47 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.653839] env[62952]: DEBUG oslo_vmware.api [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1249.653839] env[62952]: value = "task-1264019" [ 1249.653839] env[62952]: _type = "Task" [ 1249.653839] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.661668] env[62952]: DEBUG oslo_vmware.api [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264019, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.080771] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "cf523cf4-028c-4a24-9822-df384974458e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1250.081155] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "cf523cf4-028c-4a24-9822-df384974458e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1250.091790] env[62952]: DEBUG oslo_vmware.api [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1264018, 'name': PowerOffVM_Task, 'duration_secs': 0.173662} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.092577] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1250.092830] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1250.093142] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-34f71a46-c75a-4fa4-bab2-602fcd81a4b7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.148562] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1250.148799] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1250.148988] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Deleting the datastore file [datastore1] b01ea5c5-1d49-47e7-ada2-acbacbe5abdb {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1250.149268] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-875fdce2-c87e-4a8d-9ef7-7b7ee93a4169 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.155205] env[62952]: DEBUG oslo_vmware.api [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for the task: (returnval){ [ 1250.155205] env[62952]: value = "task-1264021" [ 1250.155205] env[62952]: _type = "Task" [ 1250.155205] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.165301] env[62952]: DEBUG oslo_vmware.api [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264019, 'name': ReconfigVM_Task, 'duration_secs': 0.144541} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.169259] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271995', 'volume_id': '69b2eba4-89f3-4502-ac1e-cac959b2e708', 'name': 'volume-69b2eba4-89f3-4502-ac1e-cac959b2e708', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c5d21282-1a4b-4998-8170-ee5423b958ad', 'attached_at': '', 'detached_at': '', 'volume_id': '69b2eba4-89f3-4502-ac1e-cac959b2e708', 'serial': '69b2eba4-89f3-4502-ac1e-cac959b2e708'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1250.170763] env[62952]: DEBUG oslo_vmware.api [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1264021, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.583898] env[62952]: DEBUG nova.compute.manager [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1250.668285] env[62952]: DEBUG oslo_vmware.api [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Task: {'id': task-1264021, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141561} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.668559] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1250.668745] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1250.668926] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1250.669117] env[62952]: INFO nova.compute.manager [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1250.669364] env[62952]: DEBUG oslo.service.loopingcall [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1250.669557] env[62952]: DEBUG nova.compute.manager [-] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1250.669655] env[62952]: DEBUG nova.network.neutron [-] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1251.117038] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1251.117385] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1251.118898] env[62952]: INFO nova.compute.claims [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1251.123186] env[62952]: DEBUG nova.compute.manager [req-c854dfc3-ce21-476c-ab25-6d52ca8ba842 req-493a44f8-a614-46f0-986d-12848608151d service nova] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Received event network-vif-deleted-c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1251.123373] env[62952]: INFO nova.compute.manager [req-c854dfc3-ce21-476c-ab25-6d52ca8ba842 req-493a44f8-a614-46f0-986d-12848608151d service nova] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Neutron deleted interface c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8; detaching it from the instance and deleting it from the info cache [ 1251.123545] env[62952]: DEBUG nova.network.neutron [req-c854dfc3-ce21-476c-ab25-6d52ca8ba842 req-493a44f8-a614-46f0-986d-12848608151d service nova] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1251.204558] env[62952]: DEBUG nova.objects.instance [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lazy-loading 'flavor' on Instance uuid c5d21282-1a4b-4998-8170-ee5423b958ad {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1251.598636] env[62952]: DEBUG nova.network.neutron [-] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1251.627944] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a47c2ddc-03cf-4da6-a979-e4af383b8a5d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.639665] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74865f54-4b9b-444e-8877-c9dc6ace8a81 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.662921] env[62952]: DEBUG nova.compute.manager [req-c854dfc3-ce21-476c-ab25-6d52ca8ba842 req-493a44f8-a614-46f0-986d-12848608151d service nova] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Detach interface failed, port_id=c3a25e8c-b66e-4d89-b5f1-84c7e2ae1bd8, reason: Instance b01ea5c5-1d49-47e7-ada2-acbacbe5abdb could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1251.709824] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dc9169b7-0aed-49b2-92c7-d2c75dd3b596 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "c5d21282-1a4b-4998-8170-ee5423b958ad" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.240s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1252.101484] env[62952]: INFO nova.compute.manager [-] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Took 1.43 seconds to deallocate network for instance. [ 1252.180809] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e47c8b-85ea-4c07-a753-0bc557c75ad0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.188583] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5ac72f1-105c-4a76-8b2b-bcebfd2f7397 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.220100] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75267bf0-f40a-443c-9e8d-11906b7ede0d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.227419] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac941420-8590-4069-9f08-7a5a4ff641d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.240661] env[62952]: DEBUG nova.compute.provider_tree [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1252.607944] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1252.744296] env[62952]: DEBUG nova.scheduler.client.report [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1252.850365] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "34edb671-8d37-4c08-b8ca-3b4b89b345b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1252.850640] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "34edb671-8d37-4c08-b8ca-3b4b89b345b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1253.249462] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.132s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1253.249975] env[62952]: DEBUG nova.compute.manager [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1253.252703] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.645s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1253.252936] env[62952]: DEBUG nova.objects.instance [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lazy-loading 'resources' on Instance uuid b01ea5c5-1d49-47e7-ada2-acbacbe5abdb {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1253.353451] env[62952]: DEBUG nova.compute.manager [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1253.756551] env[62952]: DEBUG nova.compute.utils [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1253.757996] env[62952]: DEBUG nova.compute.manager [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1253.758186] env[62952]: DEBUG nova.network.neutron [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1253.809166] env[62952]: DEBUG nova.policy [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '166cce5924454e73bc7deb1c77191f66', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e2f92c0076c14f15b5a4b8188f50db8e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1253.823040] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc69ccfc-3163-4327-97b9-4d23e8595338 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.830374] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-297c3a3a-9337-44b3-afa4-39276900bb0a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.865322] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76007224-2019-45cf-8253-e3e07d19cf37 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.874358] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a233a59-589d-447c-96dc-3d0ff09c4f2e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.879092] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1253.888268] env[62952]: DEBUG nova.compute.provider_tree [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1254.093749] env[62952]: DEBUG nova.network.neutron [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Successfully created port: a1a25960-ea01-416c-b893-4f824426f225 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1254.263376] env[62952]: DEBUG nova.compute.manager [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1254.391334] env[62952]: DEBUG nova.scheduler.client.report [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1254.896671] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.644s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1254.899731] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.020s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1254.903647] env[62952]: INFO nova.compute.claims [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1254.921726] env[62952]: INFO nova.scheduler.client.report [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Deleted allocations for instance b01ea5c5-1d49-47e7-ada2-acbacbe5abdb [ 1255.273183] env[62952]: DEBUG nova.compute.manager [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1255.298101] env[62952]: DEBUG nova.virt.hardware [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1255.298410] env[62952]: DEBUG nova.virt.hardware [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1255.298571] env[62952]: DEBUG nova.virt.hardware [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1255.298759] env[62952]: DEBUG nova.virt.hardware [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1255.298909] env[62952]: DEBUG nova.virt.hardware [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1255.299083] env[62952]: DEBUG nova.virt.hardware [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1255.299305] env[62952]: DEBUG nova.virt.hardware [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1255.299471] env[62952]: DEBUG nova.virt.hardware [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1255.299644] env[62952]: DEBUG nova.virt.hardware [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1255.299812] env[62952]: DEBUG nova.virt.hardware [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1255.299990] env[62952]: DEBUG nova.virt.hardware [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1255.300890] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e44ad50f-8ea3-4d5f-8488-3f5ecff74d97 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.308816] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6fd073a-2c34-4863-bcb4-2be686a0e1c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.431855] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9220bf00-96b3-4e21-89d8-30a7a4cf47b5 tempest-ServerActionsTestJSON-663735069 tempest-ServerActionsTestJSON-663735069-project-member] Lock "b01ea5c5-1d49-47e7-ada2-acbacbe5abdb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.870s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1255.454425] env[62952]: DEBUG nova.compute.manager [req-bbcd48b1-682c-4826-bae1-ccf16b67ebe5 req-743a711c-15a3-49c4-97b9-b2e41a29c0d6 service nova] [instance: cf523cf4-028c-4a24-9822-df384974458e] Received event network-vif-plugged-a1a25960-ea01-416c-b893-4f824426f225 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1255.454756] env[62952]: DEBUG oslo_concurrency.lockutils [req-bbcd48b1-682c-4826-bae1-ccf16b67ebe5 req-743a711c-15a3-49c4-97b9-b2e41a29c0d6 service nova] Acquiring lock "cf523cf4-028c-4a24-9822-df384974458e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1255.454942] env[62952]: DEBUG oslo_concurrency.lockutils [req-bbcd48b1-682c-4826-bae1-ccf16b67ebe5 req-743a711c-15a3-49c4-97b9-b2e41a29c0d6 service nova] Lock "cf523cf4-028c-4a24-9822-df384974458e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1255.455142] env[62952]: DEBUG oslo_concurrency.lockutils [req-bbcd48b1-682c-4826-bae1-ccf16b67ebe5 req-743a711c-15a3-49c4-97b9-b2e41a29c0d6 service nova] Lock "cf523cf4-028c-4a24-9822-df384974458e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1255.455353] env[62952]: DEBUG nova.compute.manager [req-bbcd48b1-682c-4826-bae1-ccf16b67ebe5 req-743a711c-15a3-49c4-97b9-b2e41a29c0d6 service nova] [instance: cf523cf4-028c-4a24-9822-df384974458e] No waiting events found dispatching network-vif-plugged-a1a25960-ea01-416c-b893-4f824426f225 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1255.455571] env[62952]: WARNING nova.compute.manager [req-bbcd48b1-682c-4826-bae1-ccf16b67ebe5 req-743a711c-15a3-49c4-97b9-b2e41a29c0d6 service nova] [instance: cf523cf4-028c-4a24-9822-df384974458e] Received unexpected event network-vif-plugged-a1a25960-ea01-416c-b893-4f824426f225 for instance with vm_state building and task_state spawning. [ 1255.553183] env[62952]: DEBUG nova.network.neutron [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Successfully updated port: a1a25960-ea01-416c-b893-4f824426f225 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1255.969087] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e662878f-eeb4-4b2c-9733-a97062877547 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.977101] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb293ef-200b-4f0c-a8bc-f0786a8e786f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.006676] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca3b512-c685-4c66-a286-ce28b7bcacc0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.013378] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b70a257-d516-4f7d-b685-e8c1200a1b25 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.025729] env[62952]: DEBUG nova.compute.provider_tree [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1256.055897] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1256.056032] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.056606] env[62952]: DEBUG nova.network.neutron [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1256.530241] env[62952]: DEBUG nova.scheduler.client.report [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1256.597764] env[62952]: DEBUG nova.network.neutron [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1256.808788] env[62952]: DEBUG nova.network.neutron [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance_info_cache with network_info: [{"id": "a1a25960-ea01-416c-b893-4f824426f225", "address": "fa:16:3e:81:97:58", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1a25960-ea", "ovs_interfaceid": "a1a25960-ea01-416c-b893-4f824426f225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1257.034569] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.136s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1257.035116] env[62952]: DEBUG nova.compute.manager [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1257.311238] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1257.311607] env[62952]: DEBUG nova.compute.manager [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Instance network_info: |[{"id": "a1a25960-ea01-416c-b893-4f824426f225", "address": "fa:16:3e:81:97:58", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1a25960-ea", "ovs_interfaceid": "a1a25960-ea01-416c-b893-4f824426f225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1257.312110] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:97:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3f499bc9-78da-46c1-9274-19edf26d31cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a1a25960-ea01-416c-b893-4f824426f225', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1257.320961] env[62952]: DEBUG oslo.service.loopingcall [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1257.321206] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf523cf4-028c-4a24-9822-df384974458e] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1257.321439] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-12f6e728-25df-4eed-840f-6b8716405545 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.341620] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1257.341620] env[62952]: value = "task-1264022" [ 1257.341620] env[62952]: _type = "Task" [ 1257.341620] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.349426] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1264022, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.486039] env[62952]: DEBUG nova.compute.manager [req-2b4d9983-9232-41c8-a25f-336bf5453898 req-3373ac26-e025-4744-84af-25636d03d4d4 service nova] [instance: cf523cf4-028c-4a24-9822-df384974458e] Received event network-changed-a1a25960-ea01-416c-b893-4f824426f225 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1257.486454] env[62952]: DEBUG nova.compute.manager [req-2b4d9983-9232-41c8-a25f-336bf5453898 req-3373ac26-e025-4744-84af-25636d03d4d4 service nova] [instance: cf523cf4-028c-4a24-9822-df384974458e] Refreshing instance network info cache due to event network-changed-a1a25960-ea01-416c-b893-4f824426f225. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1257.486603] env[62952]: DEBUG oslo_concurrency.lockutils [req-2b4d9983-9232-41c8-a25f-336bf5453898 req-3373ac26-e025-4744-84af-25636d03d4d4 service nova] Acquiring lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1257.486741] env[62952]: DEBUG oslo_concurrency.lockutils [req-2b4d9983-9232-41c8-a25f-336bf5453898 req-3373ac26-e025-4744-84af-25636d03d4d4 service nova] Acquired lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1257.486923] env[62952]: DEBUG nova.network.neutron [req-2b4d9983-9232-41c8-a25f-336bf5453898 req-3373ac26-e025-4744-84af-25636d03d4d4 service nova] [instance: cf523cf4-028c-4a24-9822-df384974458e] Refreshing network info cache for port a1a25960-ea01-416c-b893-4f824426f225 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1257.540221] env[62952]: DEBUG nova.compute.utils [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1257.541784] env[62952]: DEBUG nova.compute.manager [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1257.541784] env[62952]: DEBUG nova.network.neutron [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1257.598815] env[62952]: DEBUG nova.policy [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0a36328ec404c22adf6d2d256c9514c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f725be1feb8a44478d297b0be58c59ba', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1257.852151] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1264022, 'name': CreateVM_Task, 'duration_secs': 0.305552} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.852324] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf523cf4-028c-4a24-9822-df384974458e] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1257.852998] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1257.853187] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1257.853524] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1257.853786] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4410b1f1-edbd-49a6-9b22-f9abc41a8709 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.858337] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1257.858337] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c9b2d4-c894-8f06-1c52-16cab194b0a5" [ 1257.858337] env[62952]: _type = "Task" [ 1257.858337] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.865141] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c9b2d4-c894-8f06-1c52-16cab194b0a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.901345] env[62952]: DEBUG nova.network.neutron [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Successfully created port: 4d60af7a-a574-4e9d-a8d3-143baad5e447 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1258.044873] env[62952]: DEBUG nova.compute.manager [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1258.189612] env[62952]: DEBUG nova.network.neutron [req-2b4d9983-9232-41c8-a25f-336bf5453898 req-3373ac26-e025-4744-84af-25636d03d4d4 service nova] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updated VIF entry in instance network info cache for port a1a25960-ea01-416c-b893-4f824426f225. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1258.190209] env[62952]: DEBUG nova.network.neutron [req-2b4d9983-9232-41c8-a25f-336bf5453898 req-3373ac26-e025-4744-84af-25636d03d4d4 service nova] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance_info_cache with network_info: [{"id": "a1a25960-ea01-416c-b893-4f824426f225", "address": "fa:16:3e:81:97:58", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1a25960-ea", "ovs_interfaceid": "a1a25960-ea01-416c-b893-4f824426f225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1258.370349] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52c9b2d4-c894-8f06-1c52-16cab194b0a5, 'name': SearchDatastore_Task, 'duration_secs': 0.012323} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.370608] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1258.370925] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1258.371080] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1258.371233] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1258.371412] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1258.371671] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-549dc7cf-94d6-47ae-a43e-948a1fb2cbf2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.379728] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1258.380094] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1258.381114] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f691c06-6e75-4081-a261-2202f86b0334 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.388102] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1258.388102] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5273ab1a-ed02-0240-a0f7-fb8e46451fc2" [ 1258.388102] env[62952]: _type = "Task" [ 1258.388102] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.399515] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5273ab1a-ed02-0240-a0f7-fb8e46451fc2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.693016] env[62952]: DEBUG oslo_concurrency.lockutils [req-2b4d9983-9232-41c8-a25f-336bf5453898 req-3373ac26-e025-4744-84af-25636d03d4d4 service nova] Releasing lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1258.898994] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5273ab1a-ed02-0240-a0f7-fb8e46451fc2, 'name': SearchDatastore_Task, 'duration_secs': 0.010826} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.899876] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b023b32-d81f-4bc9-9bae-aaa1a8e05aad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.904933] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1258.904933] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52efffdf-17f2-9f67-df57-9dd76e7c2ef1" [ 1258.904933] env[62952]: _type = "Task" [ 1258.904933] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.912805] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52efffdf-17f2-9f67-df57-9dd76e7c2ef1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.053614] env[62952]: DEBUG nova.compute.manager [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1259.079807] env[62952]: DEBUG nova.virt.hardware [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1259.080090] env[62952]: DEBUG nova.virt.hardware [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1259.080262] env[62952]: DEBUG nova.virt.hardware [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1259.080446] env[62952]: DEBUG nova.virt.hardware [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1259.080596] env[62952]: DEBUG nova.virt.hardware [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1259.080751] env[62952]: DEBUG nova.virt.hardware [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1259.080955] env[62952]: DEBUG nova.virt.hardware [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1259.081140] env[62952]: DEBUG nova.virt.hardware [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1259.081407] env[62952]: DEBUG nova.virt.hardware [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1259.081584] env[62952]: DEBUG nova.virt.hardware [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1259.081760] env[62952]: DEBUG nova.virt.hardware [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1259.082628] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dccebc1-885d-4d80-9f3f-dd4babf856a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.090946] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe5ba146-fcf6-44fd-9884-ee818c81b17c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.358087] env[62952]: DEBUG nova.network.neutron [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Successfully updated port: 4d60af7a-a574-4e9d-a8d3-143baad5e447 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1259.415447] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52efffdf-17f2-9f67-df57-9dd76e7c2ef1, 'name': SearchDatastore_Task, 'duration_secs': 0.009424} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.415714] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1259.415971] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] cf523cf4-028c-4a24-9822-df384974458e/cf523cf4-028c-4a24-9822-df384974458e.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1259.416290] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a724b8a3-02ea-49e5-a7a6-3ca08ef3d31b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.422820] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1259.422820] env[62952]: value = "task-1264023" [ 1259.422820] env[62952]: _type = "Task" [ 1259.422820] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.431096] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264023, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.514050] env[62952]: DEBUG nova.compute.manager [req-6b13d352-3e51-4377-ad1d-9e93ce6a690c req-88ac0287-0043-4c28-a753-04919e950f92 service nova] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Received event network-vif-plugged-4d60af7a-a574-4e9d-a8d3-143baad5e447 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1259.514301] env[62952]: DEBUG oslo_concurrency.lockutils [req-6b13d352-3e51-4377-ad1d-9e93ce6a690c req-88ac0287-0043-4c28-a753-04919e950f92 service nova] Acquiring lock "34edb671-8d37-4c08-b8ca-3b4b89b345b9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1259.514525] env[62952]: DEBUG oslo_concurrency.lockutils [req-6b13d352-3e51-4377-ad1d-9e93ce6a690c req-88ac0287-0043-4c28-a753-04919e950f92 service nova] Lock "34edb671-8d37-4c08-b8ca-3b4b89b345b9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1259.514679] env[62952]: DEBUG oslo_concurrency.lockutils [req-6b13d352-3e51-4377-ad1d-9e93ce6a690c req-88ac0287-0043-4c28-a753-04919e950f92 service nova] Lock "34edb671-8d37-4c08-b8ca-3b4b89b345b9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1259.514849] env[62952]: DEBUG nova.compute.manager [req-6b13d352-3e51-4377-ad1d-9e93ce6a690c req-88ac0287-0043-4c28-a753-04919e950f92 service nova] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] No waiting events found dispatching network-vif-plugged-4d60af7a-a574-4e9d-a8d3-143baad5e447 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1259.515026] env[62952]: WARNING nova.compute.manager [req-6b13d352-3e51-4377-ad1d-9e93ce6a690c req-88ac0287-0043-4c28-a753-04919e950f92 service nova] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Received unexpected event network-vif-plugged-4d60af7a-a574-4e9d-a8d3-143baad5e447 for instance with vm_state building and task_state spawning. [ 1259.515195] env[62952]: DEBUG nova.compute.manager [req-6b13d352-3e51-4377-ad1d-9e93ce6a690c req-88ac0287-0043-4c28-a753-04919e950f92 service nova] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Received event network-changed-4d60af7a-a574-4e9d-a8d3-143baad5e447 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1259.515353] env[62952]: DEBUG nova.compute.manager [req-6b13d352-3e51-4377-ad1d-9e93ce6a690c req-88ac0287-0043-4c28-a753-04919e950f92 service nova] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Refreshing instance network info cache due to event network-changed-4d60af7a-a574-4e9d-a8d3-143baad5e447. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1259.515537] env[62952]: DEBUG oslo_concurrency.lockutils [req-6b13d352-3e51-4377-ad1d-9e93ce6a690c req-88ac0287-0043-4c28-a753-04919e950f92 service nova] Acquiring lock "refresh_cache-34edb671-8d37-4c08-b8ca-3b4b89b345b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1259.515674] env[62952]: DEBUG oslo_concurrency.lockutils [req-6b13d352-3e51-4377-ad1d-9e93ce6a690c req-88ac0287-0043-4c28-a753-04919e950f92 service nova] Acquired lock "refresh_cache-34edb671-8d37-4c08-b8ca-3b4b89b345b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1259.515833] env[62952]: DEBUG nova.network.neutron [req-6b13d352-3e51-4377-ad1d-9e93ce6a690c req-88ac0287-0043-4c28-a753-04919e950f92 service nova] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Refreshing network info cache for port 4d60af7a-a574-4e9d-a8d3-143baad5e447 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1259.861136] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "refresh_cache-34edb671-8d37-4c08-b8ca-3b4b89b345b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1259.932334] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264023, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.43568} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.932606] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] cf523cf4-028c-4a24-9822-df384974458e/cf523cf4-028c-4a24-9822-df384974458e.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1259.932823] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1259.933085] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd99a3a7-1cb6-4f0a-8c4e-71183717f8e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.940007] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1259.940007] env[62952]: value = "task-1264024" [ 1259.940007] env[62952]: _type = "Task" [ 1259.940007] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.947802] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264024, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.047323] env[62952]: DEBUG nova.network.neutron [req-6b13d352-3e51-4377-ad1d-9e93ce6a690c req-88ac0287-0043-4c28-a753-04919e950f92 service nova] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1260.117124] env[62952]: DEBUG nova.network.neutron [req-6b13d352-3e51-4377-ad1d-9e93ce6a690c req-88ac0287-0043-4c28-a753-04919e950f92 service nova] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1260.450307] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264024, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055647} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.450676] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1260.451523] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d776a81-d545-4b89-abef-be284a1a1dca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.472699] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] cf523cf4-028c-4a24-9822-df384974458e/cf523cf4-028c-4a24-9822-df384974458e.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1260.472974] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-215526fa-8bce-44d0-88c6-4dcafcbe7ec6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.491644] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1260.491644] env[62952]: value = "task-1264025" [ 1260.491644] env[62952]: _type = "Task" [ 1260.491644] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.500294] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264025, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.619764] env[62952]: DEBUG oslo_concurrency.lockutils [req-6b13d352-3e51-4377-ad1d-9e93ce6a690c req-88ac0287-0043-4c28-a753-04919e950f92 service nova] Releasing lock "refresh_cache-34edb671-8d37-4c08-b8ca-3b4b89b345b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1260.620208] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquired lock "refresh_cache-34edb671-8d37-4c08-b8ca-3b4b89b345b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1260.620344] env[62952]: DEBUG nova.network.neutron [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1261.001813] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264025, 'name': ReconfigVM_Task, 'duration_secs': 0.256935} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.002179] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Reconfigured VM instance instance-0000006c to attach disk [datastore2] cf523cf4-028c-4a24-9822-df384974458e/cf523cf4-028c-4a24-9822-df384974458e.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1261.002758] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1a4531b2-c644-4e3d-b55f-21618e8f2924 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.008776] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1261.008776] env[62952]: value = "task-1264026" [ 1261.008776] env[62952]: _type = "Task" [ 1261.008776] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.016113] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264026, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.150225] env[62952]: DEBUG nova.network.neutron [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1261.265593] env[62952]: DEBUG nova.network.neutron [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Updating instance_info_cache with network_info: [{"id": "4d60af7a-a574-4e9d-a8d3-143baad5e447", "address": "fa:16:3e:c9:2d:c7", "network": {"id": "8e167a8b-2628-43e1-8418-e673064745b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-232555460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f725be1feb8a44478d297b0be58c59ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d60af7a-a5", "ovs_interfaceid": "4d60af7a-a574-4e9d-a8d3-143baad5e447", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1261.518598] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264026, 'name': Rename_Task, 'duration_secs': 0.130407} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.518857] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1261.519112] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f69e31d-085b-4485-930d-7dc42854c757 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.525017] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1261.525017] env[62952]: value = "task-1264027" [ 1261.525017] env[62952]: _type = "Task" [ 1261.525017] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.531899] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264027, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.768677] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Releasing lock "refresh_cache-34edb671-8d37-4c08-b8ca-3b4b89b345b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1261.769022] env[62952]: DEBUG nova.compute.manager [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Instance network_info: |[{"id": "4d60af7a-a574-4e9d-a8d3-143baad5e447", "address": "fa:16:3e:c9:2d:c7", "network": {"id": "8e167a8b-2628-43e1-8418-e673064745b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-232555460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f725be1feb8a44478d297b0be58c59ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d60af7a-a5", "ovs_interfaceid": "4d60af7a-a574-4e9d-a8d3-143baad5e447", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1261.769461] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:2d:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69351262-8d39-441a-85ba-3a78df436d17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4d60af7a-a574-4e9d-a8d3-143baad5e447', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1261.776857] env[62952]: DEBUG oslo.service.loopingcall [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1261.777094] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1261.777325] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9283cd9f-b0b0-4331-b21b-67fe7aaf522d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.796879] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1261.796879] env[62952]: value = "task-1264028" [ 1261.796879] env[62952]: _type = "Task" [ 1261.796879] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.803891] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1264028, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.034783] env[62952]: DEBUG oslo_vmware.api [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264027, 'name': PowerOnVM_Task, 'duration_secs': 0.412229} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.035098] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1262.035269] env[62952]: INFO nova.compute.manager [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Took 6.76 seconds to spawn the instance on the hypervisor. [ 1262.035449] env[62952]: DEBUG nova.compute.manager [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1262.036234] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841b3cc5-652b-40b3-aee9-b55db6ee14d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.307903] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1264028, 'name': CreateVM_Task, 'duration_secs': 0.284718} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.308100] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1262.308655] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1262.308817] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1262.309156] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1262.309413] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-856fc1a3-1657-49b8-9b1b-62531b7bdceb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.313857] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1262.313857] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525015e2-6ba5-3dbd-edf1-455f30204216" [ 1262.313857] env[62952]: _type = "Task" [ 1262.313857] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.320812] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525015e2-6ba5-3dbd-edf1-455f30204216, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.555711] env[62952]: INFO nova.compute.manager [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Took 11.47 seconds to build instance. [ 1262.825134] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]525015e2-6ba5-3dbd-edf1-455f30204216, 'name': SearchDatastore_Task, 'duration_secs': 0.008818} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.825475] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1262.825710] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1262.825946] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1262.826120] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1262.826336] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1262.826607] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62d5a562-e3ae-4476-88b0-8c850d1e6d47 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.835560] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1262.835741] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1262.836481] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce305768-729b-47dc-9bf2-fb5cdb8039e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.841285] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1262.841285] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52321f75-22e0-9669-a41e-fcd465c75220" [ 1262.841285] env[62952]: _type = "Task" [ 1262.841285] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.848383] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52321f75-22e0-9669-a41e-fcd465c75220, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.058356] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d61a2579-9660-44fb-bf6d-6741c05715f9 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "cf523cf4-028c-4a24-9822-df384974458e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.977s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1263.351156] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52321f75-22e0-9669-a41e-fcd465c75220, 'name': SearchDatastore_Task, 'duration_secs': 0.007748} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.351948] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ed98d4b-993d-4bba-b587-31cf45c04965 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.356849] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1263.356849] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5287cd26-f325-425d-7a48-d6c0993a286c" [ 1263.356849] env[62952]: _type = "Task" [ 1263.356849] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.364164] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5287cd26-f325-425d-7a48-d6c0993a286c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.404598] env[62952]: DEBUG nova.compute.manager [req-5c61f5ce-b527-4ee0-a38d-c4af9001f0b7 req-7718eaac-6f92-4c3e-8ab5-5e0471dcdcc8 service nova] [instance: cf523cf4-028c-4a24-9822-df384974458e] Received event network-changed-a1a25960-ea01-416c-b893-4f824426f225 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1263.404806] env[62952]: DEBUG nova.compute.manager [req-5c61f5ce-b527-4ee0-a38d-c4af9001f0b7 req-7718eaac-6f92-4c3e-8ab5-5e0471dcdcc8 service nova] [instance: cf523cf4-028c-4a24-9822-df384974458e] Refreshing instance network info cache due to event network-changed-a1a25960-ea01-416c-b893-4f824426f225. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1263.405033] env[62952]: DEBUG oslo_concurrency.lockutils [req-5c61f5ce-b527-4ee0-a38d-c4af9001f0b7 req-7718eaac-6f92-4c3e-8ab5-5e0471dcdcc8 service nova] Acquiring lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1263.405181] env[62952]: DEBUG oslo_concurrency.lockutils [req-5c61f5ce-b527-4ee0-a38d-c4af9001f0b7 req-7718eaac-6f92-4c3e-8ab5-5e0471dcdcc8 service nova] Acquired lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1263.405350] env[62952]: DEBUG nova.network.neutron [req-5c61f5ce-b527-4ee0-a38d-c4af9001f0b7 req-7718eaac-6f92-4c3e-8ab5-5e0471dcdcc8 service nova] [instance: cf523cf4-028c-4a24-9822-df384974458e] Refreshing network info cache for port a1a25960-ea01-416c-b893-4f824426f225 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1263.867409] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5287cd26-f325-425d-7a48-d6c0993a286c, 'name': SearchDatastore_Task, 'duration_secs': 0.008326} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.867700] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1263.867924] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 34edb671-8d37-4c08-b8ca-3b4b89b345b9/34edb671-8d37-4c08-b8ca-3b4b89b345b9.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1263.868181] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f771a421-e92b-48b3-8ba7-e9809b5f4fcf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.875219] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1263.875219] env[62952]: value = "task-1264029" [ 1263.875219] env[62952]: _type = "Task" [ 1263.875219] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.882864] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264029, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.125774] env[62952]: DEBUG nova.network.neutron [req-5c61f5ce-b527-4ee0-a38d-c4af9001f0b7 req-7718eaac-6f92-4c3e-8ab5-5e0471dcdcc8 service nova] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updated VIF entry in instance network info cache for port a1a25960-ea01-416c-b893-4f824426f225. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1264.126248] env[62952]: DEBUG nova.network.neutron [req-5c61f5ce-b527-4ee0-a38d-c4af9001f0b7 req-7718eaac-6f92-4c3e-8ab5-5e0471dcdcc8 service nova] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance_info_cache with network_info: [{"id": "a1a25960-ea01-416c-b893-4f824426f225", "address": "fa:16:3e:81:97:58", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1a25960-ea", "ovs_interfaceid": "a1a25960-ea01-416c-b893-4f824426f225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1264.385274] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264029, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.434758} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.385546] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 34edb671-8d37-4c08-b8ca-3b4b89b345b9/34edb671-8d37-4c08-b8ca-3b4b89b345b9.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1264.385793] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1264.386021] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da189d42-a0ec-4215-b32b-e4670207190b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.392476] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1264.392476] env[62952]: value = "task-1264030" [ 1264.392476] env[62952]: _type = "Task" [ 1264.392476] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.400472] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264030, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.629449] env[62952]: DEBUG oslo_concurrency.lockutils [req-5c61f5ce-b527-4ee0-a38d-c4af9001f0b7 req-7718eaac-6f92-4c3e-8ab5-5e0471dcdcc8 service nova] Releasing lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1264.903099] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264030, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.053775} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.903099] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1264.903319] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4f3a5a-5bef-4507-b0c0-6ba259b87a4d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.925644] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 34edb671-8d37-4c08-b8ca-3b4b89b345b9/34edb671-8d37-4c08-b8ca-3b4b89b345b9.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1264.925892] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53f480f4-0e73-4fad-8dd5-43f39101772f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.944856] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1264.944856] env[62952]: value = "task-1264031" [ 1264.944856] env[62952]: _type = "Task" [ 1264.944856] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.952612] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264031, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.454710] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264031, 'name': ReconfigVM_Task, 'duration_secs': 0.263965} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.455054] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 34edb671-8d37-4c08-b8ca-3b4b89b345b9/34edb671-8d37-4c08-b8ca-3b4b89b345b9.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1265.455610] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c8619af7-49e1-4f77-9f5d-d005c021466b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.461979] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1265.461979] env[62952]: value = "task-1264032" [ 1265.461979] env[62952]: _type = "Task" [ 1265.461979] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.469165] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264032, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.971971] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264032, 'name': Rename_Task, 'duration_secs': 0.192029} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.972270] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1265.972518] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73079234-7951-400c-9627-a1cba404d8b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.979225] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1265.979225] env[62952]: value = "task-1264033" [ 1265.979225] env[62952]: _type = "Task" [ 1265.979225] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.986314] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264033, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.488813] env[62952]: DEBUG oslo_vmware.api [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264033, 'name': PowerOnVM_Task, 'duration_secs': 0.431816} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.489120] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1266.489307] env[62952]: INFO nova.compute.manager [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Took 7.44 seconds to spawn the instance on the hypervisor. [ 1266.489495] env[62952]: DEBUG nova.compute.manager [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1266.490290] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43b342e-4b21-4c0c-a89b-42a50ea98dbb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.007056] env[62952]: INFO nova.compute.manager [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Took 13.15 seconds to build instance. [ 1267.509179] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42f5ef5e-8dd9-40e3-8c21-50dede95d84c tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "34edb671-8d37-4c08-b8ca-3b4b89b345b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.658s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1268.077284] env[62952]: DEBUG nova.compute.manager [req-8b3c7023-19d0-47ee-9b84-914b1a0018a5 req-35846a74-d7f8-41b5-8b42-db44a5872a7a service nova] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Received event network-changed-4d60af7a-a574-4e9d-a8d3-143baad5e447 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1268.077603] env[62952]: DEBUG nova.compute.manager [req-8b3c7023-19d0-47ee-9b84-914b1a0018a5 req-35846a74-d7f8-41b5-8b42-db44a5872a7a service nova] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Refreshing instance network info cache due to event network-changed-4d60af7a-a574-4e9d-a8d3-143baad5e447. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1268.077923] env[62952]: DEBUG oslo_concurrency.lockutils [req-8b3c7023-19d0-47ee-9b84-914b1a0018a5 req-35846a74-d7f8-41b5-8b42-db44a5872a7a service nova] Acquiring lock "refresh_cache-34edb671-8d37-4c08-b8ca-3b4b89b345b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1268.078174] env[62952]: DEBUG oslo_concurrency.lockutils [req-8b3c7023-19d0-47ee-9b84-914b1a0018a5 req-35846a74-d7f8-41b5-8b42-db44a5872a7a service nova] Acquired lock "refresh_cache-34edb671-8d37-4c08-b8ca-3b4b89b345b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1268.078443] env[62952]: DEBUG nova.network.neutron [req-8b3c7023-19d0-47ee-9b84-914b1a0018a5 req-35846a74-d7f8-41b5-8b42-db44a5872a7a service nova] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Refreshing network info cache for port 4d60af7a-a574-4e9d-a8d3-143baad5e447 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1268.271855] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1268.778601] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1268.778945] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1268.778997] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1268.779158] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1268.780057] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0905f054-7160-430d-85ed-9bd653274ff5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.789849] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-507a4f11-38f0-4165-9d68-7ec97b63093a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.794399] env[62952]: DEBUG nova.network.neutron [req-8b3c7023-19d0-47ee-9b84-914b1a0018a5 req-35846a74-d7f8-41b5-8b42-db44a5872a7a service nova] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Updated VIF entry in instance network info cache for port 4d60af7a-a574-4e9d-a8d3-143baad5e447. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1268.794731] env[62952]: DEBUG nova.network.neutron [req-8b3c7023-19d0-47ee-9b84-914b1a0018a5 req-35846a74-d7f8-41b5-8b42-db44a5872a7a service nova] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Updating instance_info_cache with network_info: [{"id": "4d60af7a-a574-4e9d-a8d3-143baad5e447", "address": "fa:16:3e:c9:2d:c7", "network": {"id": "8e167a8b-2628-43e1-8418-e673064745b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-232555460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f725be1feb8a44478d297b0be58c59ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d60af7a-a5", "ovs_interfaceid": "4d60af7a-a574-4e9d-a8d3-143baad5e447", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1268.807523] env[62952]: DEBUG oslo_concurrency.lockutils [req-8b3c7023-19d0-47ee-9b84-914b1a0018a5 req-35846a74-d7f8-41b5-8b42-db44a5872a7a service nova] Releasing lock "refresh_cache-34edb671-8d37-4c08-b8ca-3b4b89b345b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1268.808364] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b594673a-3205-4046-a905-1ccccea0eb07 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.815303] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce458b04-7381-4074-9181-8edc62d700a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.844554] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181174MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1268.844729] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1268.844902] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.872869] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance c5d21282-1a4b-4998-8170-ee5423b958ad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1269.873307] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance cf523cf4-028c-4a24-9822-df384974458e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1269.873307] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 34edb671-8d37-4c08-b8ca-3b4b89b345b9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1269.873391] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1269.873497] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1269.921634] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cffa99c7-5393-4b9d-b56b-42e9bb037e60 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.928988] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-effe4464-93d7-4ad7-9b17-6139b824b027 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.962349] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc053b8f-75b5-46b7-b90e-ba8c22cc3231 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.970241] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e76964-5685-4355-8d4c-4cce782551c0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.983357] env[62952]: DEBUG nova.compute.provider_tree [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1270.486552] env[62952]: DEBUG nova.scheduler.client.report [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1270.991203] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1270.991638] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.146s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.992999] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1271.993379] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1273.035407] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "refresh_cache-c5d21282-1a4b-4998-8170-ee5423b958ad" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1273.035670] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquired lock "refresh_cache-c5d21282-1a4b-4998-8170-ee5423b958ad" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1273.035735] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Forcefully refreshing network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1274.252852] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Updating instance_info_cache with network_info: [{"id": "f5eb1e66-2290-4179-b630-173dd8f3f058", "address": "fa:16:3e:15:7b:1a", "network": {"id": "8e167a8b-2628-43e1-8418-e673064745b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-232555460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f725be1feb8a44478d297b0be58c59ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5eb1e66-22", "ovs_interfaceid": "f5eb1e66-2290-4179-b630-173dd8f3f058", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1274.756200] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Releasing lock "refresh_cache-c5d21282-1a4b-4998-8170-ee5423b958ad" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1274.756502] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Updated the network info_cache for instance {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1274.756737] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1274.756879] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1274.757038] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1274.757195] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1274.757342] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1274.757489] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1274.757617] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1278.033495] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1278.033842] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1300.571357] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "cf523cf4-028c-4a24-9822-df384974458e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1300.571650] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "cf523cf4-028c-4a24-9822-df384974458e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1301.074685] env[62952]: DEBUG nova.compute.utils [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1301.578402] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "cf523cf4-028c-4a24-9822-df384974458e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1302.643977] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "cf523cf4-028c-4a24-9822-df384974458e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1302.644419] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "cf523cf4-028c-4a24-9822-df384974458e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1302.644521] env[62952]: INFO nova.compute.manager [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Attaching volume 4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6 to /dev/sdb [ 1302.674212] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60c79e92-a35f-4d47-b3db-9dbf62c8d846 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.681422] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75f680ec-7fb9-49c7-99ff-933e697cc066 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.694351] env[62952]: DEBUG nova.virt.block_device [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating existing volume attachment record: 8d10b1c0-4dab-4c5e-af6f-22ac71d3bb3a {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1304.533397] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "34edb671-8d37-4c08-b8ca-3b4b89b345b9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1304.533777] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "34edb671-8d37-4c08-b8ca-3b4b89b345b9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1304.533878] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "34edb671-8d37-4c08-b8ca-3b4b89b345b9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1304.534076] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "34edb671-8d37-4c08-b8ca-3b4b89b345b9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1304.534256] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "34edb671-8d37-4c08-b8ca-3b4b89b345b9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1304.537601] env[62952]: INFO nova.compute.manager [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Terminating instance [ 1304.539405] env[62952]: DEBUG nova.compute.manager [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1304.539644] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1304.540455] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c547da3-ad7f-4eee-b0f2-db7c904948ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.549388] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1304.549605] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cfe786d1-84e5-48ae-8f6c-7b690eb7a6fe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.555541] env[62952]: DEBUG oslo_vmware.api [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1304.555541] env[62952]: value = "task-1264037" [ 1304.555541] env[62952]: _type = "Task" [ 1304.555541] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.562871] env[62952]: DEBUG oslo_vmware.api [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264037, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.065321] env[62952]: DEBUG oslo_vmware.api [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264037, 'name': PowerOffVM_Task, 'duration_secs': 0.180501} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.065595] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1305.065758] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1305.066011] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a793ec8a-2956-419c-9c13-f336ef8e46e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.128197] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1305.128405] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1305.128599] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Deleting the datastore file [datastore2] 34edb671-8d37-4c08-b8ca-3b4b89b345b9 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1305.128856] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-64e99423-5a06-4e20-876f-c23ca1005d9e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.135641] env[62952]: DEBUG oslo_vmware.api [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1305.135641] env[62952]: value = "task-1264040" [ 1305.135641] env[62952]: _type = "Task" [ 1305.135641] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1305.142856] env[62952]: DEBUG oslo_vmware.api [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264040, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.645685] env[62952]: DEBUG oslo_vmware.api [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264040, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133872} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.645685] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1305.645685] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1305.646140] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1305.646140] env[62952]: INFO nova.compute.manager [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1305.646257] env[62952]: DEBUG oslo.service.loopingcall [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1305.646458] env[62952]: DEBUG nova.compute.manager [-] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1305.646553] env[62952]: DEBUG nova.network.neutron [-] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1306.063161] env[62952]: DEBUG nova.compute.manager [req-c856ac1f-cf44-4677-808e-c81ce9a491e0 req-323e60fa-b1e4-42ef-90d6-dc6921576207 service nova] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Received event network-vif-deleted-4d60af7a-a574-4e9d-a8d3-143baad5e447 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1306.063432] env[62952]: INFO nova.compute.manager [req-c856ac1f-cf44-4677-808e-c81ce9a491e0 req-323e60fa-b1e4-42ef-90d6-dc6921576207 service nova] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Neutron deleted interface 4d60af7a-a574-4e9d-a8d3-143baad5e447; detaching it from the instance and deleting it from the info cache [ 1306.063543] env[62952]: DEBUG nova.network.neutron [req-c856ac1f-cf44-4677-808e-c81ce9a491e0 req-323e60fa-b1e4-42ef-90d6-dc6921576207 service nova] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1306.535584] env[62952]: DEBUG nova.network.neutron [-] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1306.566007] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-afa4b2ac-51bd-4a81-859c-7cd55a2efe6f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.575843] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36d06ef-3b18-4a55-9afb-812442792942 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.600885] env[62952]: DEBUG nova.compute.manager [req-c856ac1f-cf44-4677-808e-c81ce9a491e0 req-323e60fa-b1e4-42ef-90d6-dc6921576207 service nova] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Detach interface failed, port_id=4d60af7a-a574-4e9d-a8d3-143baad5e447, reason: Instance 34edb671-8d37-4c08-b8ca-3b4b89b345b9 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1307.039128] env[62952]: INFO nova.compute.manager [-] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Took 1.39 seconds to deallocate network for instance. [ 1307.236996] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1307.237269] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271999', 'volume_id': '4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6', 'name': 'volume-4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cf523cf4-028c-4a24-9822-df384974458e', 'attached_at': '', 'detached_at': '', 'volume_id': '4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6', 'serial': '4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1307.238174] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fa971ee-1377-4b23-81d9-a0065cacdd2d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.255567] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d6643b-ab25-4604-b590-76a83d3c4d21 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.279493] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] volume-4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6/volume-4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1307.279725] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-30753bac-35f0-4daf-812c-038445f91509 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.297103] env[62952]: DEBUG oslo_vmware.api [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1307.297103] env[62952]: value = "task-1264041" [ 1307.297103] env[62952]: _type = "Task" [ 1307.297103] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.304528] env[62952]: DEBUG oslo_vmware.api [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264041, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.545379] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1307.545626] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1307.545896] env[62952]: DEBUG nova.objects.instance [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lazy-loading 'resources' on Instance uuid 34edb671-8d37-4c08-b8ca-3b4b89b345b9 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1307.806598] env[62952]: DEBUG oslo_vmware.api [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264041, 'name': ReconfigVM_Task, 'duration_secs': 0.314064} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1307.806922] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Reconfigured VM instance instance-0000006c to attach disk [datastore1] volume-4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6/volume-4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1307.811778] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb4c1867-6883-49e5-a71e-5afbd8a4c619 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.826843] env[62952]: DEBUG oslo_vmware.api [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1307.826843] env[62952]: value = "task-1264042" [ 1307.826843] env[62952]: _type = "Task" [ 1307.826843] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.834238] env[62952]: DEBUG oslo_vmware.api [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264042, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.101724] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5a3c556-fcb8-427b-b5f4-d4cdfc9e9fff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.109054] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5604265-c566-46d0-b657-c42ebbb9003e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.138159] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-110bb58f-fd1f-4f19-b464-cbb7e1ccb0e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.145047] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b0bd23-f4fb-4369-bdfa-4476d4367633 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.159088] env[62952]: DEBUG nova.compute.provider_tree [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1308.336302] env[62952]: DEBUG oslo_vmware.api [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264042, 'name': ReconfigVM_Task, 'duration_secs': 0.128925} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.336603] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271999', 'volume_id': '4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6', 'name': 'volume-4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cf523cf4-028c-4a24-9822-df384974458e', 'attached_at': '', 'detached_at': '', 'volume_id': '4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6', 'serial': '4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1308.662333] env[62952]: DEBUG nova.scheduler.client.report [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1309.167355] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.622s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1309.187901] env[62952]: INFO nova.scheduler.client.report [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Deleted allocations for instance 34edb671-8d37-4c08-b8ca-3b4b89b345b9 [ 1309.370737] env[62952]: DEBUG nova.objects.instance [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lazy-loading 'flavor' on Instance uuid cf523cf4-028c-4a24-9822-df384974458e {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1309.695740] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e48b6e11-0965-4b65-9aef-4147ab005251 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "34edb671-8d37-4c08-b8ca-3b4b89b345b9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.162s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1309.875951] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75c6b79f-795d-4465-9d29-f68ea9d888ec tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "cf523cf4-028c-4a24-9822-df384974458e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.232s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1310.461471] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "c5d21282-1a4b-4998-8170-ee5423b958ad" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1310.461860] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "c5d21282-1a4b-4998-8170-ee5423b958ad" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1310.834016] env[62952]: DEBUG nova.compute.manager [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Stashing vm_state: active {{(pid=62952) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1310.965182] env[62952]: INFO nova.compute.manager [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Detaching volume 69b2eba4-89f3-4502-ac1e-cac959b2e708 [ 1310.998739] env[62952]: INFO nova.virt.block_device [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Attempting to driver detach volume 69b2eba4-89f3-4502-ac1e-cac959b2e708 from mountpoint /dev/sdb [ 1310.998989] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1310.999226] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271995', 'volume_id': '69b2eba4-89f3-4502-ac1e-cac959b2e708', 'name': 'volume-69b2eba4-89f3-4502-ac1e-cac959b2e708', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c5d21282-1a4b-4998-8170-ee5423b958ad', 'attached_at': '', 'detached_at': '', 'volume_id': '69b2eba4-89f3-4502-ac1e-cac959b2e708', 'serial': '69b2eba4-89f3-4502-ac1e-cac959b2e708'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1311.000108] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba997fc-83c9-49c1-8e8c-a639d2f51625 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.021438] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e67ba4-0fa8-4b3f-956d-14f6d26cd1cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.028291] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-424fb24b-9b0b-4bcf-bdc4-ed50e96a79ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.047577] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7084e0-4d4d-4e9d-abb0-f68c06c5f6e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.061499] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] The volume has not been displaced from its original location: [datastore2] volume-69b2eba4-89f3-4502-ac1e-cac959b2e708/volume-69b2eba4-89f3-4502-ac1e-cac959b2e708.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1311.066599] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Reconfiguring VM instance instance-0000006b to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1311.066894] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10034e73-4a84-46a6-b50e-3ffcc4c5df33 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.083687] env[62952]: DEBUG oslo_vmware.api [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1311.083687] env[62952]: value = "task-1264043" [ 1311.083687] env[62952]: _type = "Task" [ 1311.083687] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1311.091020] env[62952]: DEBUG oslo_vmware.api [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264043, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.350657] env[62952]: DEBUG oslo_concurrency.lockutils [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1311.350920] env[62952]: DEBUG oslo_concurrency.lockutils [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1311.593276] env[62952]: DEBUG oslo_vmware.api [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264043, 'name': ReconfigVM_Task, 'duration_secs': 0.228426} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1311.593601] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Reconfigured VM instance instance-0000006b to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1311.598047] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56a42142-2753-45cd-be42-4f4c2b4da3d6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.612282] env[62952]: DEBUG oslo_vmware.api [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1311.612282] env[62952]: value = "task-1264044" [ 1311.612282] env[62952]: _type = "Task" [ 1311.612282] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1311.619497] env[62952]: DEBUG oslo_vmware.api [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264044, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.856062] env[62952]: INFO nova.compute.claims [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1312.124266] env[62952]: DEBUG oslo_vmware.api [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264044, 'name': ReconfigVM_Task, 'duration_secs': 0.123567} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1312.124563] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271995', 'volume_id': '69b2eba4-89f3-4502-ac1e-cac959b2e708', 'name': 'volume-69b2eba4-89f3-4502-ac1e-cac959b2e708', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c5d21282-1a4b-4998-8170-ee5423b958ad', 'attached_at': '', 'detached_at': '', 'volume_id': '69b2eba4-89f3-4502-ac1e-cac959b2e708', 'serial': '69b2eba4-89f3-4502-ac1e-cac959b2e708'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1312.362091] env[62952]: INFO nova.compute.resource_tracker [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating resource usage from migration 63d1d248-1f5a-4f12-bf3b-4162bf7ee0e9 [ 1312.407566] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be4978e-6614-4bb9-a73c-2f0f6ebb5e1a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.414915] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a1ef7a-ab93-48bd-bfa0-600bae6e9c70 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.443776] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513d7d72-97b6-4016-a460-f70517a8e2f1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.450756] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0cfa40f-976f-4f63-b249-e5fa7c3c9786 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.463419] env[62952]: DEBUG nova.compute.provider_tree [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1312.664139] env[62952]: DEBUG nova.objects.instance [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lazy-loading 'flavor' on Instance uuid c5d21282-1a4b-4998-8170-ee5423b958ad {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1312.966609] env[62952]: DEBUG nova.scheduler.client.report [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1313.471679] env[62952]: DEBUG oslo_concurrency.lockutils [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.121s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1313.471942] env[62952]: INFO nova.compute.manager [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Migrating [ 1313.670892] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1787411-e3d8-4acc-bbba-e6bc2bae14c9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "c5d21282-1a4b-4998-8170-ee5423b958ad" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.209s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1313.987655] env[62952]: DEBUG oslo_concurrency.lockutils [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1313.987871] env[62952]: DEBUG oslo_concurrency.lockutils [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1313.988052] env[62952]: DEBUG nova.network.neutron [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1314.687943] env[62952]: DEBUG nova.network.neutron [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance_info_cache with network_info: [{"id": "a1a25960-ea01-416c-b893-4f824426f225", "address": "fa:16:3e:81:97:58", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1a25960-ea", "ovs_interfaceid": "a1a25960-ea01-416c-b893-4f824426f225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1314.692861] env[62952]: DEBUG oslo_concurrency.lockutils [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "c5d21282-1a4b-4998-8170-ee5423b958ad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1314.693098] env[62952]: DEBUG oslo_concurrency.lockutils [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "c5d21282-1a4b-4998-8170-ee5423b958ad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1314.693301] env[62952]: DEBUG oslo_concurrency.lockutils [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "c5d21282-1a4b-4998-8170-ee5423b958ad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1314.693482] env[62952]: DEBUG oslo_concurrency.lockutils [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "c5d21282-1a4b-4998-8170-ee5423b958ad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1314.693653] env[62952]: DEBUG oslo_concurrency.lockutils [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "c5d21282-1a4b-4998-8170-ee5423b958ad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1314.695531] env[62952]: INFO nova.compute.manager [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Terminating instance [ 1314.697459] env[62952]: DEBUG nova.compute.manager [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1314.697654] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1314.698507] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8688d103-ffb6-4a77-9101-56741ee4718b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.706358] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1314.706573] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4cb4508f-c83c-4c5e-814e-e9995a6ab80b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.712367] env[62952]: DEBUG oslo_vmware.api [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1314.712367] env[62952]: value = "task-1264045" [ 1314.712367] env[62952]: _type = "Task" [ 1314.712367] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1314.719681] env[62952]: DEBUG oslo_vmware.api [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264045, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1315.191196] env[62952]: DEBUG oslo_concurrency.lockutils [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1315.222387] env[62952]: DEBUG oslo_vmware.api [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264045, 'name': PowerOffVM_Task, 'duration_secs': 0.182651} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1315.222658] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1315.222829] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1315.223103] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e6a2a090-8f8a-4fae-9323-78751b6b12de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.287188] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1315.287402] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1315.287584] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Deleting the datastore file [datastore1] c5d21282-1a4b-4998-8170-ee5423b958ad {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1315.287874] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf8a32f6-eec5-4b24-94d1-76fecc102b16 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.293784] env[62952]: DEBUG oslo_vmware.api [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1315.293784] env[62952]: value = "task-1264047" [ 1315.293784] env[62952]: _type = "Task" [ 1315.293784] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1315.301214] env[62952]: DEBUG oslo_vmware.api [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264047, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1315.802962] env[62952]: DEBUG oslo_vmware.api [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264047, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140199} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1315.803374] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1315.803465] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1315.803650] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1315.803827] env[62952]: INFO nova.compute.manager [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1315.804143] env[62952]: DEBUG oslo.service.loopingcall [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1315.804354] env[62952]: DEBUG nova.compute.manager [-] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1315.804455] env[62952]: DEBUG nova.network.neutron [-] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1316.205631] env[62952]: DEBUG nova.compute.manager [req-6b7c0f44-4d8a-41ab-a7b2-e02cf1ba72d2 req-4ea19ec6-44e4-4c93-80a5-74828ae1bd7b service nova] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Received event network-vif-deleted-f5eb1e66-2290-4179-b630-173dd8f3f058 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1316.206711] env[62952]: INFO nova.compute.manager [req-6b7c0f44-4d8a-41ab-a7b2-e02cf1ba72d2 req-4ea19ec6-44e4-4c93-80a5-74828ae1bd7b service nova] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Neutron deleted interface f5eb1e66-2290-4179-b630-173dd8f3f058; detaching it from the instance and deleting it from the info cache [ 1316.206711] env[62952]: DEBUG nova.network.neutron [req-6b7c0f44-4d8a-41ab-a7b2-e02cf1ba72d2 req-4ea19ec6-44e4-4c93-80a5-74828ae1bd7b service nova] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1316.686503] env[62952]: DEBUG nova.network.neutron [-] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1316.705634] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e01002-d0b5-424a-ae52-62b5382eff2c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.708896] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-87416409-59f3-4a42-86b1-1c772ce0e714 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.729318] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance 'cf523cf4-028c-4a24-9822-df384974458e' progress to 0 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1316.736333] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4f89ef5-d30a-4d8f-b0ad-9485c185b5eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.760193] env[62952]: DEBUG nova.compute.manager [req-6b7c0f44-4d8a-41ab-a7b2-e02cf1ba72d2 req-4ea19ec6-44e4-4c93-80a5-74828ae1bd7b service nova] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Detach interface failed, port_id=f5eb1e66-2290-4179-b630-173dd8f3f058, reason: Instance c5d21282-1a4b-4998-8170-ee5423b958ad could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1317.189843] env[62952]: INFO nova.compute.manager [-] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Took 1.39 seconds to deallocate network for instance. [ 1317.249050] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1317.249678] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e2783310-e0b1-47a0-8dbd-c24b7b8d4a4a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.256300] env[62952]: DEBUG oslo_vmware.api [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1317.256300] env[62952]: value = "task-1264048" [ 1317.256300] env[62952]: _type = "Task" [ 1317.256300] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1317.264227] env[62952]: DEBUG oslo_vmware.api [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264048, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1317.697728] env[62952]: DEBUG oslo_concurrency.lockutils [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1317.697728] env[62952]: DEBUG oslo_concurrency.lockutils [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1317.697728] env[62952]: DEBUG nova.objects.instance [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lazy-loading 'resources' on Instance uuid c5d21282-1a4b-4998-8170-ee5423b958ad {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1317.765245] env[62952]: DEBUG oslo_vmware.api [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264048, 'name': PowerOffVM_Task, 'duration_secs': 0.185931} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1317.765472] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1317.765659] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance 'cf523cf4-028c-4a24-9822-df384974458e' progress to 17 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1318.252366] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf6336ec-b1ec-4fac-82c1-bec03aec639b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.261283] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249911a2-200f-4901-abee-cb5611bf38c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.290909] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1318.291163] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1318.291327] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1318.291511] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1318.291661] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1318.291810] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1318.292032] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1318.292207] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1318.292376] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1318.292538] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1318.292709] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1318.297657] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84971c1f-e3bf-4722-b37b-caad12949d20 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.308031] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981fb2ca-7544-490d-a16e-b24b1e1b840f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.316773] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2cd933-d84b-43b3-adac-aef9af24ee28 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.320798] env[62952]: DEBUG oslo_vmware.api [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1318.320798] env[62952]: value = "task-1264049" [ 1318.320798] env[62952]: _type = "Task" [ 1318.320798] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1318.331279] env[62952]: DEBUG nova.compute.provider_tree [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1318.337039] env[62952]: DEBUG oslo_vmware.api [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1318.830653] env[62952]: DEBUG oslo_vmware.api [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264049, 'name': ReconfigVM_Task, 'duration_secs': 0.166008} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1318.830956] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance 'cf523cf4-028c-4a24-9822-df384974458e' progress to 33 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1318.834946] env[62952]: DEBUG nova.scheduler.client.report [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1319.340530] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1319.340975] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1319.340975] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1319.341131] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1319.341284] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1319.341436] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1319.341643] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1319.341808] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1319.341976] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1319.342163] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1319.342340] env[62952]: DEBUG nova.virt.hardware [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1319.347846] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1319.348524] env[62952]: DEBUG oslo_concurrency.lockutils [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.651s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1319.350673] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-11f0c124-f5a0-4297-92a6-ed464728781a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.365758] env[62952]: INFO nova.scheduler.client.report [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Deleted allocations for instance c5d21282-1a4b-4998-8170-ee5423b958ad [ 1319.371804] env[62952]: DEBUG oslo_vmware.api [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1319.371804] env[62952]: value = "task-1264050" [ 1319.371804] env[62952]: _type = "Task" [ 1319.371804] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1319.381190] env[62952]: DEBUG oslo_vmware.api [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264050, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.878781] env[62952]: DEBUG oslo_concurrency.lockutils [None req-26aaccd9-94bc-4328-8cc0-3487fb6a6a90 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "c5d21282-1a4b-4998-8170-ee5423b958ad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.186s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1319.885537] env[62952]: DEBUG oslo_vmware.api [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264050, 'name': ReconfigVM_Task, 'duration_secs': 0.167888} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1319.885916] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1319.886669] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be24a36e-601f-437e-b2d3-30b97b0c504b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.914029] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] cf523cf4-028c-4a24-9822-df384974458e/cf523cf4-028c-4a24-9822-df384974458e.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1319.915207] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f304083-5a28-4ddf-ab09-aeed6a7ae816 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.939063] env[62952]: DEBUG oslo_vmware.api [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1319.939063] env[62952]: value = "task-1264051" [ 1319.939063] env[62952]: _type = "Task" [ 1319.939063] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1319.947570] env[62952]: DEBUG oslo_vmware.api [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264051, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.450563] env[62952]: DEBUG oslo_vmware.api [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264051, 'name': ReconfigVM_Task, 'duration_secs': 0.280226} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.450907] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Reconfigured VM instance instance-0000006c to attach disk [datastore2] cf523cf4-028c-4a24-9822-df384974458e/cf523cf4-028c-4a24-9822-df384974458e.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1320.450997] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance 'cf523cf4-028c-4a24-9822-df384974458e' progress to 50 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1320.957639] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac554a73-9a2d-410c-b787-cc972025f4f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.980132] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378057ac-a38d-4c0c-ae79-4ff1047fd14c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.999875] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance 'cf523cf4-028c-4a24-9822-df384974458e' progress to 67 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1321.588030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1321.588289] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1322.093639] env[62952]: DEBUG nova.compute.manager [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1322.617439] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1322.617713] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1322.619378] env[62952]: INFO nova.compute.claims [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1322.639823] env[62952]: DEBUG nova.network.neutron [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Port a1a25960-ea01-416c-b893-4f824426f225 binding to destination host cpu-1 is already ACTIVE {{(pid=62952) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1323.661763] env[62952]: DEBUG oslo_concurrency.lockutils [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "cf523cf4-028c-4a24-9822-df384974458e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1323.662151] env[62952]: DEBUG oslo_concurrency.lockutils [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "cf523cf4-028c-4a24-9822-df384974458e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1323.662447] env[62952]: DEBUG oslo_concurrency.lockutils [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "cf523cf4-028c-4a24-9822-df384974458e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1323.693184] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f222744e-720f-4778-b089-1cea47f2dcb9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.700714] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c7b324-ec8b-4b25-9cf9-547ff65f470e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.732430] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b3864c-cbb8-408f-82ed-4e0be9715a0a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.739863] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21471ff9-0d65-4ee8-a16d-f3f9e0a2473d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.752880] env[62952]: DEBUG nova.compute.provider_tree [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1324.255688] env[62952]: DEBUG nova.scheduler.client.report [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1324.697212] env[62952]: DEBUG oslo_concurrency.lockutils [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1324.697508] env[62952]: DEBUG oslo_concurrency.lockutils [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1324.697582] env[62952]: DEBUG nova.network.neutron [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1324.760316] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.142s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1324.760817] env[62952]: DEBUG nova.compute.manager [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1325.266063] env[62952]: DEBUG nova.compute.utils [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1325.267740] env[62952]: DEBUG nova.compute.manager [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1325.267881] env[62952]: DEBUG nova.network.neutron [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1325.305661] env[62952]: DEBUG nova.policy [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0a36328ec404c22adf6d2d256c9514c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f725be1feb8a44478d297b0be58c59ba', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1325.432012] env[62952]: DEBUG nova.network.neutron [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance_info_cache with network_info: [{"id": "a1a25960-ea01-416c-b893-4f824426f225", "address": "fa:16:3e:81:97:58", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1a25960-ea", "ovs_interfaceid": "a1a25960-ea01-416c-b893-4f824426f225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1325.564692] env[62952]: DEBUG nova.network.neutron [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Successfully created port: ed11c649-dae7-4ce6-95cd-5ef951865824 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1325.771088] env[62952]: DEBUG nova.compute.manager [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1325.935579] env[62952]: DEBUG oslo_concurrency.lockutils [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1326.445712] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dc42d78-f7c3-4b67-81b9-acd2a2be2c4e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.453987] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d45663f-4c7d-4a3b-8021-69d70e5d0eed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.780491] env[62952]: DEBUG nova.compute.manager [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1326.805660] env[62952]: DEBUG nova.virt.hardware [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1326.805926] env[62952]: DEBUG nova.virt.hardware [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1326.806098] env[62952]: DEBUG nova.virt.hardware [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1326.806288] env[62952]: DEBUG nova.virt.hardware [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1326.806438] env[62952]: DEBUG nova.virt.hardware [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1326.806590] env[62952]: DEBUG nova.virt.hardware [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1326.806800] env[62952]: DEBUG nova.virt.hardware [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1326.806988] env[62952]: DEBUG nova.virt.hardware [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1326.807186] env[62952]: DEBUG nova.virt.hardware [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1326.807353] env[62952]: DEBUG nova.virt.hardware [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1326.807526] env[62952]: DEBUG nova.virt.hardware [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1326.808414] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-730451c5-86f5-431a-bf89-0f348c226f15 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.816366] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2527163b-cae9-42b1-8e3d-310b51eba3f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.965363] env[62952]: DEBUG nova.compute.manager [req-e4714979-7a87-4a06-9735-464c18520309 req-8626b397-cf05-40e8-a38c-0c5afdf00b94 service nova] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Received event network-vif-plugged-ed11c649-dae7-4ce6-95cd-5ef951865824 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1326.965492] env[62952]: DEBUG oslo_concurrency.lockutils [req-e4714979-7a87-4a06-9735-464c18520309 req-8626b397-cf05-40e8-a38c-0c5afdf00b94 service nova] Acquiring lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1326.965712] env[62952]: DEBUG oslo_concurrency.lockutils [req-e4714979-7a87-4a06-9735-464c18520309 req-8626b397-cf05-40e8-a38c-0c5afdf00b94 service nova] Lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1326.965841] env[62952]: DEBUG oslo_concurrency.lockutils [req-e4714979-7a87-4a06-9735-464c18520309 req-8626b397-cf05-40e8-a38c-0c5afdf00b94 service nova] Lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1326.966026] env[62952]: DEBUG nova.compute.manager [req-e4714979-7a87-4a06-9735-464c18520309 req-8626b397-cf05-40e8-a38c-0c5afdf00b94 service nova] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] No waiting events found dispatching network-vif-plugged-ed11c649-dae7-4ce6-95cd-5ef951865824 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1326.966204] env[62952]: WARNING nova.compute.manager [req-e4714979-7a87-4a06-9735-464c18520309 req-8626b397-cf05-40e8-a38c-0c5afdf00b94 service nova] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Received unexpected event network-vif-plugged-ed11c649-dae7-4ce6-95cd-5ef951865824 for instance with vm_state building and task_state spawning. [ 1327.471246] env[62952]: DEBUG nova.network.neutron [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Successfully updated port: ed11c649-dae7-4ce6-95cd-5ef951865824 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1327.515215] env[62952]: DEBUG nova.compute.manager [req-5199b92c-0270-480b-a55f-bf6ad0f9b6aa req-508dd55f-7cae-4265-b251-781fa79f9d75 service nova] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Received event network-changed-ed11c649-dae7-4ce6-95cd-5ef951865824 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1327.515215] env[62952]: DEBUG nova.compute.manager [req-5199b92c-0270-480b-a55f-bf6ad0f9b6aa req-508dd55f-7cae-4265-b251-781fa79f9d75 service nova] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Refreshing instance network info cache due to event network-changed-ed11c649-dae7-4ce6-95cd-5ef951865824. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1327.515302] env[62952]: DEBUG oslo_concurrency.lockutils [req-5199b92c-0270-480b-a55f-bf6ad0f9b6aa req-508dd55f-7cae-4265-b251-781fa79f9d75 service nova] Acquiring lock "refresh_cache-b5da3312-a52c-4941-9c3c-247f4d93fbbf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1327.515424] env[62952]: DEBUG oslo_concurrency.lockutils [req-5199b92c-0270-480b-a55f-bf6ad0f9b6aa req-508dd55f-7cae-4265-b251-781fa79f9d75 service nova] Acquired lock "refresh_cache-b5da3312-a52c-4941-9c3c-247f4d93fbbf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1327.515588] env[62952]: DEBUG nova.network.neutron [req-5199b92c-0270-480b-a55f-bf6ad0f9b6aa req-508dd55f-7cae-4265-b251-781fa79f9d75 service nova] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Refreshing network info cache for port ed11c649-dae7-4ce6-95cd-5ef951865824 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1327.548344] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc46bd0e-8651-4aef-ad12-058f9cfbdfe8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.569975] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb78a5e-6ed3-4c79-9569-d8dd6bf1144e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.576563] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance 'cf523cf4-028c-4a24-9822-df384974458e' progress to 83 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1327.974956] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "refresh_cache-b5da3312-a52c-4941-9c3c-247f4d93fbbf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1328.046213] env[62952]: DEBUG nova.network.neutron [req-5199b92c-0270-480b-a55f-bf6ad0f9b6aa req-508dd55f-7cae-4265-b251-781fa79f9d75 service nova] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1328.083605] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1328.083900] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b0a025cc-8079-4a87-ba91-b4fb7e13af6a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.091267] env[62952]: DEBUG oslo_vmware.api [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1328.091267] env[62952]: value = "task-1264052" [ 1328.091267] env[62952]: _type = "Task" [ 1328.091267] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.098928] env[62952]: DEBUG oslo_vmware.api [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264052, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.117107] env[62952]: DEBUG nova.network.neutron [req-5199b92c-0270-480b-a55f-bf6ad0f9b6aa req-508dd55f-7cae-4265-b251-781fa79f9d75 service nova] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1328.601317] env[62952]: DEBUG oslo_vmware.api [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264052, 'name': PowerOnVM_Task, 'duration_secs': 0.373234} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.601591] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1328.601782] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-333fc0b3-60b1-4534-81f1-bece7a8b4753 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance 'cf523cf4-028c-4a24-9822-df384974458e' progress to 100 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1328.619101] env[62952]: DEBUG oslo_concurrency.lockutils [req-5199b92c-0270-480b-a55f-bf6ad0f9b6aa req-508dd55f-7cae-4265-b251-781fa79f9d75 service nova] Releasing lock "refresh_cache-b5da3312-a52c-4941-9c3c-247f4d93fbbf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1328.619457] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquired lock "refresh_cache-b5da3312-a52c-4941-9c3c-247f4d93fbbf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1328.619614] env[62952]: DEBUG nova.network.neutron [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1329.150414] env[62952]: DEBUG nova.network.neutron [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1329.272053] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.272229] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1329.272351] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1329.280504] env[62952]: DEBUG nova.network.neutron [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Updating instance_info_cache with network_info: [{"id": "ed11c649-dae7-4ce6-95cd-5ef951865824", "address": "fa:16:3e:01:99:6d", "network": {"id": "8e167a8b-2628-43e1-8418-e673064745b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-232555460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f725be1feb8a44478d297b0be58c59ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped11c649-da", "ovs_interfaceid": "ed11c649-dae7-4ce6-95cd-5ef951865824", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1329.776182] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1329.776450] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1329.776645] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquired lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1329.776827] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: cf523cf4-028c-4a24-9822-df384974458e] Forcefully refreshing network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1329.777045] env[62952]: DEBUG nova.objects.instance [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lazy-loading 'info_cache' on Instance uuid cf523cf4-028c-4a24-9822-df384974458e {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1329.783042] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Releasing lock "refresh_cache-b5da3312-a52c-4941-9c3c-247f4d93fbbf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1329.783318] env[62952]: DEBUG nova.compute.manager [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Instance network_info: |[{"id": "ed11c649-dae7-4ce6-95cd-5ef951865824", "address": "fa:16:3e:01:99:6d", "network": {"id": "8e167a8b-2628-43e1-8418-e673064745b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-232555460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f725be1feb8a44478d297b0be58c59ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped11c649-da", "ovs_interfaceid": "ed11c649-dae7-4ce6-95cd-5ef951865824", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1329.783740] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:99:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69351262-8d39-441a-85ba-3a78df436d17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ed11c649-dae7-4ce6-95cd-5ef951865824', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1329.791703] env[62952]: DEBUG oslo.service.loopingcall [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1329.792515] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1329.793139] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c9188c7e-4e85-4e53-adf4-ae9a9950ec38 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.812811] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1329.812811] env[62952]: value = "task-1264053" [ 1329.812811] env[62952]: _type = "Task" [ 1329.812811] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1329.820481] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1264053, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.322549] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1264053, 'name': CreateVM_Task, 'duration_secs': 0.291662} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.322878] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1330.323332] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1330.323504] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1330.323837] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1330.324096] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-609a51dd-37ab-4f8c-9795-1f59945c9a86 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.328471] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1330.328471] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e155bc-46a2-75f9-03be-100f5e70321e" [ 1330.328471] env[62952]: _type = "Task" [ 1330.328471] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.335801] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e155bc-46a2-75f9-03be-100f5e70321e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.641570] env[62952]: DEBUG nova.network.neutron [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Port a1a25960-ea01-416c-b893-4f824426f225 binding to destination host cpu-1 is already ACTIVE {{(pid=62952) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1330.641828] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1330.840762] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e155bc-46a2-75f9-03be-100f5e70321e, 'name': SearchDatastore_Task, 'duration_secs': 0.012045} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.841133] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1330.841423] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1330.841623] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1330.841771] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1330.841958] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1330.842239] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2e6be3df-c0e3-4345-9492-21eae81fcd8d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.850848] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1330.851232] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1330.851954] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6dc85149-4e2d-452e-b928-e6adf61c30ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.857861] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1330.857861] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a234b6-7c9f-da63-f6f6-4eb04543fcec" [ 1330.857861] env[62952]: _type = "Task" [ 1330.857861] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.864986] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a234b6-7c9f-da63-f6f6-4eb04543fcec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.367659] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a234b6-7c9f-da63-f6f6-4eb04543fcec, 'name': SearchDatastore_Task, 'duration_secs': 0.009365} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.368456] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3535d166-9acd-4ceb-aa8b-e2d1bbf80021 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.373599] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1331.373599] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5287aac0-0166-e208-2fea-79ea4b3ca32c" [ 1331.373599] env[62952]: _type = "Task" [ 1331.373599] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.381138] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5287aac0-0166-e208-2fea-79ea4b3ca32c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.496483] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance_info_cache with network_info: [{"id": "a1a25960-ea01-416c-b893-4f824426f225", "address": "fa:16:3e:81:97:58", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1a25960-ea", "ovs_interfaceid": "a1a25960-ea01-416c-b893-4f824426f225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1331.883794] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5287aac0-0166-e208-2fea-79ea4b3ca32c, 'name': SearchDatastore_Task, 'duration_secs': 0.00965} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.884118] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1331.884390] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] b5da3312-a52c-4941-9c3c-247f4d93fbbf/b5da3312-a52c-4941-9c3c-247f4d93fbbf.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1331.884652] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-04a38503-9d83-49a3-b35b-ca2da333fe78 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.891716] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1331.891716] env[62952]: value = "task-1264054" [ 1331.891716] env[62952]: _type = "Task" [ 1331.891716] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.898673] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264054, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.999620] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Releasing lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1331.999846] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updated the network info_cache for instance {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1332.000155] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1332.000342] env[62952]: DEBUG nova.network.neutron [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1332.001614] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1332.001792] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1332.001958] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1332.002626] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1332.003026] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1332.402049] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264054, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468077} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1332.402340] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] b5da3312-a52c-4941-9c3c-247f4d93fbbf/b5da3312-a52c-4941-9c3c-247f4d93fbbf.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1332.402559] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1332.402814] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-62ff4c80-22e3-46f2-b93e-3e106fb17631 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.410045] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1332.410045] env[62952]: value = "task-1264055" [ 1332.410045] env[62952]: _type = "Task" [ 1332.410045] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1332.419835] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264055, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.506325] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1332.506503] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1332.506664] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1332.506814] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1332.508067] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea75cba-fcbe-4ecf-b1de-bf04b10387d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.516058] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e95ad197-04ba-4685-8d93-a8350c664196 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.529664] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a30a3a67-9e71-463c-b008-afc1733b2100 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.537617] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95528b7-99dc-40fd-8c6d-7ac10c9f209e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.566138] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180991MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1332.566290] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1332.566468] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1332.701533] env[62952]: DEBUG nova.network.neutron [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance_info_cache with network_info: [{"id": "a1a25960-ea01-416c-b893-4f824426f225", "address": "fa:16:3e:81:97:58", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1a25960-ea", "ovs_interfaceid": "a1a25960-ea01-416c-b893-4f824426f225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1332.920141] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264055, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080034} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1332.920362] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1332.921124] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b1cab6-0e0c-439d-b7d1-5dc3c4f8ca89 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.941990] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] b5da3312-a52c-4941-9c3c-247f4d93fbbf/b5da3312-a52c-4941-9c3c-247f4d93fbbf.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1332.942230] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-26416d9e-a887-44f2-aa3f-201b03d81fe9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.960952] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1332.960952] env[62952]: value = "task-1264056" [ 1332.960952] env[62952]: _type = "Task" [ 1332.960952] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1332.968048] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264056, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.204500] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1333.470025] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264056, 'name': ReconfigVM_Task, 'duration_secs': 0.29049} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1333.470409] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Reconfigured VM instance instance-0000006e to attach disk [datastore1] b5da3312-a52c-4941-9c3c-247f4d93fbbf/b5da3312-a52c-4941-9c3c-247f4d93fbbf.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1333.470912] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d0ab0bfc-b0fd-4526-9658-e7d917ed8bed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.476859] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1333.476859] env[62952]: value = "task-1264057" [ 1333.476859] env[62952]: _type = "Task" [ 1333.476859] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1333.484411] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264057, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.574571] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Applying migration context for instance cf523cf4-028c-4a24-9822-df384974458e as it has an incoming, in-progress migration 63d1d248-1f5a-4f12-bf3b-4162bf7ee0e9. Migration status is reverting {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1333.575218] env[62952]: INFO nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating resource usage from migration 63d1d248-1f5a-4f12-bf3b-4162bf7ee0e9 [ 1333.596853] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Migration 63d1d248-1f5a-4f12-bf3b-4162bf7ee0e9 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1333.597009] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance cf523cf4-028c-4a24-9822-df384974458e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1333.597190] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance b5da3312-a52c-4941-9c3c-247f4d93fbbf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1333.597370] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1333.597944] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1152MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1333.642804] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af5f226c-78d9-4eda-bcce-7d213e7c0222 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.649644] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac374f0-689b-43cb-91a9-927e95c4591f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.677962] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8778332b-22c3-4b16-822b-711ad6eb5ea7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.684302] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee06ac6-a298-4a5e-b530-31ee7381d055 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.696394] env[62952]: DEBUG nova.compute.provider_tree [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1333.708141] env[62952]: DEBUG nova.compute.manager [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62952) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1333.986689] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264057, 'name': Rename_Task, 'duration_secs': 0.136458} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1333.986958] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1333.987250] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a07f12a4-54f7-4fe4-859b-bbfff38dde9b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.992989] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1333.992989] env[62952]: value = "task-1264058" [ 1333.992989] env[62952]: _type = "Task" [ 1333.992989] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.001201] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264058, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.199066] env[62952]: DEBUG nova.scheduler.client.report [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1334.502575] env[62952]: DEBUG oslo_vmware.api [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264058, 'name': PowerOnVM_Task, 'duration_secs': 0.449025} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.503049] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1334.503049] env[62952]: INFO nova.compute.manager [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Took 7.72 seconds to spawn the instance on the hypervisor. [ 1334.503245] env[62952]: DEBUG nova.compute.manager [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1334.503990] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9b4c9d-c641-43dd-ad82-0d324d190a7b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.704498] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1334.704739] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.138s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1334.811068] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1334.811068] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1334.975392] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1334.975666] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1335.020961] env[62952]: INFO nova.compute.manager [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Took 12.42 seconds to build instance. [ 1335.313637] env[62952]: DEBUG nova.objects.instance [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lazy-loading 'migration_context' on Instance uuid cf523cf4-028c-4a24-9822-df384974458e {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1335.522806] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4897a037-806f-4524-9cda-c2550bc3bae1 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.934s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1335.860807] env[62952]: DEBUG nova.compute.manager [req-da0ed631-08ab-47b7-8d05-13ceae908b5e req-893c2a40-aab5-419a-8c2c-6e3b8c0cc50c service nova] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Received event network-changed-ed11c649-dae7-4ce6-95cd-5ef951865824 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1335.860807] env[62952]: DEBUG nova.compute.manager [req-da0ed631-08ab-47b7-8d05-13ceae908b5e req-893c2a40-aab5-419a-8c2c-6e3b8c0cc50c service nova] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Refreshing instance network info cache due to event network-changed-ed11c649-dae7-4ce6-95cd-5ef951865824. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1335.861140] env[62952]: DEBUG oslo_concurrency.lockutils [req-da0ed631-08ab-47b7-8d05-13ceae908b5e req-893c2a40-aab5-419a-8c2c-6e3b8c0cc50c service nova] Acquiring lock "refresh_cache-b5da3312-a52c-4941-9c3c-247f4d93fbbf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1335.861196] env[62952]: DEBUG oslo_concurrency.lockutils [req-da0ed631-08ab-47b7-8d05-13ceae908b5e req-893c2a40-aab5-419a-8c2c-6e3b8c0cc50c service nova] Acquired lock "refresh_cache-b5da3312-a52c-4941-9c3c-247f4d93fbbf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1335.861366] env[62952]: DEBUG nova.network.neutron [req-da0ed631-08ab-47b7-8d05-13ceae908b5e req-893c2a40-aab5-419a-8c2c-6e3b8c0cc50c service nova] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Refreshing network info cache for port ed11c649-dae7-4ce6-95cd-5ef951865824 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1335.875264] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d013a04-72e9-418d-aef3-d68c7f19d13d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.883610] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4b8f1f6-2fc5-48b4-b95c-12aa317824e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.915560] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c6407e-7160-42ce-98f9-d568faa23278 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.923184] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8939e4-0fa1-4fd1-86fd-a1fcd8ba97a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.937146] env[62952]: DEBUG nova.compute.provider_tree [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1336.272358] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1336.439958] env[62952]: DEBUG nova.scheduler.client.report [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1336.566480] env[62952]: DEBUG nova.network.neutron [req-da0ed631-08ab-47b7-8d05-13ceae908b5e req-893c2a40-aab5-419a-8c2c-6e3b8c0cc50c service nova] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Updated VIF entry in instance network info cache for port ed11c649-dae7-4ce6-95cd-5ef951865824. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1336.566852] env[62952]: DEBUG nova.network.neutron [req-da0ed631-08ab-47b7-8d05-13ceae908b5e req-893c2a40-aab5-419a-8c2c-6e3b8c0cc50c service nova] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Updating instance_info_cache with network_info: [{"id": "ed11c649-dae7-4ce6-95cd-5ef951865824", "address": "fa:16:3e:01:99:6d", "network": {"id": "8e167a8b-2628-43e1-8418-e673064745b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-232555460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.226", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f725be1feb8a44478d297b0be58c59ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped11c649-da", "ovs_interfaceid": "ed11c649-dae7-4ce6-95cd-5ef951865824", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1337.069892] env[62952]: DEBUG oslo_concurrency.lockutils [req-da0ed631-08ab-47b7-8d05-13ceae908b5e req-893c2a40-aab5-419a-8c2c-6e3b8c0cc50c service nova] Releasing lock "refresh_cache-b5da3312-a52c-4941-9c3c-247f4d93fbbf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1337.267931] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1337.451225] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.641s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1338.986788] env[62952]: INFO nova.compute.manager [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Swapping old allocation on dict_keys(['615127b5-dc82-4491-9c8f-4e6d0caa0690']) held by migration 63d1d248-1f5a-4f12-bf3b-4162bf7ee0e9 for instance [ 1339.008057] env[62952]: DEBUG nova.scheduler.client.report [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Overwriting current allocation {'allocations': {'615127b5-dc82-4491-9c8f-4e6d0caa0690': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 147}}, 'project_id': 'e2f92c0076c14f15b5a4b8188f50db8e', 'user_id': '166cce5924454e73bc7deb1c77191f66', 'consumer_generation': 1} on consumer cf523cf4-028c-4a24-9822-df384974458e {{(pid=62952) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1339.080100] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1339.080334] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1339.080539] env[62952]: DEBUG nova.network.neutron [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1339.779874] env[62952]: DEBUG nova.network.neutron [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance_info_cache with network_info: [{"id": "a1a25960-ea01-416c-b893-4f824426f225", "address": "fa:16:3e:81:97:58", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1a25960-ea", "ovs_interfaceid": "a1a25960-ea01-416c-b893-4f824426f225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1340.282792] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "refresh_cache-cf523cf4-028c-4a24-9822-df384974458e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1340.283833] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be995361-b3f8-4f9b-aa4f-21c5d3382243 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.291183] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2caa40db-58b8-4d80-b7f7-97f589f085e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.369807] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1341.370244] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-56b26b7c-acfa-4fa7-85c2-56053784158c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.377797] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1341.377797] env[62952]: value = "task-1264059" [ 1341.377797] env[62952]: _type = "Task" [ 1341.377797] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.385929] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264059, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.887660] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264059, 'name': PowerOffVM_Task, 'duration_secs': 0.183926} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.887907] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1341.888570] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1341.888778] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1341.888936] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1341.889142] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1341.889302] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1341.889461] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1341.889665] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1341.889833] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1341.889995] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1341.890179] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1341.890357] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1341.895178] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd66fc6a-2679-4016-a4a5-467ebb69dc40 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.910835] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1341.910835] env[62952]: value = "task-1264060" [ 1341.910835] env[62952]: _type = "Task" [ 1341.910835] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.918335] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264060, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.421027] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264060, 'name': ReconfigVM_Task, 'duration_secs': 0.151581} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1342.421888] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c77b80de-b227-47d7-8694-76325b854bd3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.445313] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1342.445606] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1342.445770] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1342.445958] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1342.446127] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1342.446282] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1342.446498] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1342.446655] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1342.446823] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1342.446988] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1342.447187] env[62952]: DEBUG nova.virt.hardware [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1342.448009] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-660074a3-aacc-4ddb-af56-5f03a506669f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.453885] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1342.453885] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52966228-b833-4dd1-ce61-7198a8bbc186" [ 1342.453885] env[62952]: _type = "Task" [ 1342.453885] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1342.461567] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52966228-b833-4dd1-ce61-7198a8bbc186, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.963911] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52966228-b833-4dd1-ce61-7198a8bbc186, 'name': SearchDatastore_Task, 'duration_secs': 0.017861} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1342.969074] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1342.969349] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ea028ba-17f2-49e4-b9c3-e74007e7485c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.986593] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1342.986593] env[62952]: value = "task-1264061" [ 1342.986593] env[62952]: _type = "Task" [ 1342.986593] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1342.993910] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264061, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1343.495908] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264061, 'name': ReconfigVM_Task, 'duration_secs': 0.20267} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1343.496290] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1343.497147] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8d8205-c1a2-489b-b6b9-3c01d718a75a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.521531] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] cf523cf4-028c-4a24-9822-df384974458e/cf523cf4-028c-4a24-9822-df384974458e.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1343.521799] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06a8a8cb-a564-4e61-a41c-91c2a4942439 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.539337] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1343.539337] env[62952]: value = "task-1264062" [ 1343.539337] env[62952]: _type = "Task" [ 1343.539337] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1343.546680] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264062, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1344.048952] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264062, 'name': ReconfigVM_Task, 'duration_secs': 0.278457} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1344.049254] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Reconfigured VM instance instance-0000006c to attach disk [datastore2] cf523cf4-028c-4a24-9822-df384974458e/cf523cf4-028c-4a24-9822-df384974458e.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1344.050068] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf6d460-63d3-4ca0-ba88-5fc5e6102d1f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.070039] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd091a37-6cb5-450d-ba29-e5450939007f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.089846] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1bad304-ce62-4724-9ece-526d6896183b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.109172] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4215a880-ffc3-43dd-91a2-888215814cdd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.115134] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1344.115348] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-80702304-c53f-4575-9464-31ea378a0c08 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.120376] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1344.120376] env[62952]: value = "task-1264063" [ 1344.120376] env[62952]: _type = "Task" [ 1344.120376] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.127015] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264063, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1344.630400] env[62952]: DEBUG oslo_vmware.api [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264063, 'name': PowerOnVM_Task, 'duration_secs': 0.413492} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1344.630767] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1345.671616] env[62952]: INFO nova.compute.manager [None req-3849497c-1408-4bdb-a320-b755890c6792 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance to original state: 'active' [ 1347.618488] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "cf523cf4-028c-4a24-9822-df384974458e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1347.618819] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "cf523cf4-028c-4a24-9822-df384974458e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1347.619023] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "cf523cf4-028c-4a24-9822-df384974458e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1347.619212] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "cf523cf4-028c-4a24-9822-df384974458e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1347.619392] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "cf523cf4-028c-4a24-9822-df384974458e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1347.621660] env[62952]: INFO nova.compute.manager [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Terminating instance [ 1347.623606] env[62952]: DEBUG nova.compute.manager [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1347.623832] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1347.624087] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bbda9e11-c234-4ec1-b1a5-a854afa7928c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.630867] env[62952]: DEBUG oslo_vmware.api [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1347.630867] env[62952]: value = "task-1264064" [ 1347.630867] env[62952]: _type = "Task" [ 1347.630867] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.639238] env[62952]: DEBUG oslo_vmware.api [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264064, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1348.140582] env[62952]: DEBUG oslo_vmware.api [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264064, 'name': PowerOffVM_Task, 'duration_secs': 0.198071} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1348.140848] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1348.140992] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1348.141209] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271999', 'volume_id': '4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6', 'name': 'volume-4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'cf523cf4-028c-4a24-9822-df384974458e', 'attached_at': '2024-10-10T21:43:13.000000', 'detached_at': '', 'volume_id': '4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6', 'serial': '4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1348.141948] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-407da6ce-bc66-4137-ae91-f0c8f444c923 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.162445] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2d7d78-fe63-4148-aafd-23d33a61c418 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.168726] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed48f74-c35d-4397-8f55-d8ae80ede150 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.188951] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83c5fc1-63cc-40ee-baf1-440a1e68a337 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.202757] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] The volume has not been displaced from its original location: [datastore1] volume-4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6/volume-4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1348.207885] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Reconfiguring VM instance instance-0000006c to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1348.208157] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-361255a1-29dd-4c09-b9bf-e4e8b64f2b44 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.225108] env[62952]: DEBUG oslo_vmware.api [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1348.225108] env[62952]: value = "task-1264065" [ 1348.225108] env[62952]: _type = "Task" [ 1348.225108] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1348.232500] env[62952]: DEBUG oslo_vmware.api [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264065, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1348.734405] env[62952]: DEBUG oslo_vmware.api [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264065, 'name': ReconfigVM_Task, 'duration_secs': 0.187918} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1348.734732] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Reconfigured VM instance instance-0000006c to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1348.739362] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e57d8c0-b571-43be-a995-a68ff00f5ff0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.753917] env[62952]: DEBUG oslo_vmware.api [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1348.753917] env[62952]: value = "task-1264066" [ 1348.753917] env[62952]: _type = "Task" [ 1348.753917] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1348.761922] env[62952]: DEBUG oslo_vmware.api [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264066, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1349.263392] env[62952]: DEBUG oslo_vmware.api [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264066, 'name': ReconfigVM_Task, 'duration_secs': 0.12497} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1349.263646] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-271999', 'volume_id': '4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6', 'name': 'volume-4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'cf523cf4-028c-4a24-9822-df384974458e', 'attached_at': '2024-10-10T21:43:13.000000', 'detached_at': '', 'volume_id': '4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6', 'serial': '4e6a0d93-567d-41cf-b2c5-603ac3cbfaa6'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1349.263933] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1349.264711] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49b6c6a3-0cbd-4822-a1f6-c51fb77b00ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.271459] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1349.271702] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d91523e-9690-47a7-bbfa-25eb52c60636 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.337736] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1349.337958] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1349.338215] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Deleting the datastore file [datastore2] cf523cf4-028c-4a24-9822-df384974458e {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1349.338553] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b46b4695-236c-4cb3-ae01-0e4d84784a9e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.344722] env[62952]: DEBUG oslo_vmware.api [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1349.344722] env[62952]: value = "task-1264068" [ 1349.344722] env[62952]: _type = "Task" [ 1349.344722] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1349.352562] env[62952]: DEBUG oslo_vmware.api [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264068, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1349.855704] env[62952]: DEBUG oslo_vmware.api [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264068, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158186} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1349.856089] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1349.856157] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1349.856349] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1349.856526] env[62952]: INFO nova.compute.manager [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Took 2.23 seconds to destroy the instance on the hypervisor. [ 1349.856775] env[62952]: DEBUG oslo.service.loopingcall [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1349.856975] env[62952]: DEBUG nova.compute.manager [-] [instance: cf523cf4-028c-4a24-9822-df384974458e] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1349.857084] env[62952]: DEBUG nova.network.neutron [-] [instance: cf523cf4-028c-4a24-9822-df384974458e] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1350.329940] env[62952]: DEBUG nova.compute.manager [req-7ad70d67-3625-4f1a-a69b-7bf448db87e7 req-b260120a-c302-4b81-be2f-4068826f96ad service nova] [instance: cf523cf4-028c-4a24-9822-df384974458e] Received event network-vif-deleted-a1a25960-ea01-416c-b893-4f824426f225 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1350.330415] env[62952]: INFO nova.compute.manager [req-7ad70d67-3625-4f1a-a69b-7bf448db87e7 req-b260120a-c302-4b81-be2f-4068826f96ad service nova] [instance: cf523cf4-028c-4a24-9822-df384974458e] Neutron deleted interface a1a25960-ea01-416c-b893-4f824426f225; detaching it from the instance and deleting it from the info cache [ 1350.330611] env[62952]: DEBUG nova.network.neutron [req-7ad70d67-3625-4f1a-a69b-7bf448db87e7 req-b260120a-c302-4b81-be2f-4068826f96ad service nova] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1350.783357] env[62952]: DEBUG nova.network.neutron [-] [instance: cf523cf4-028c-4a24-9822-df384974458e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1350.833181] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6a59dcb4-854f-4e0e-b776-73cd6bed5763 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.842539] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e898811-b869-4efd-806f-b83f3d59e5dc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.866300] env[62952]: DEBUG nova.compute.manager [req-7ad70d67-3625-4f1a-a69b-7bf448db87e7 req-b260120a-c302-4b81-be2f-4068826f96ad service nova] [instance: cf523cf4-028c-4a24-9822-df384974458e] Detach interface failed, port_id=a1a25960-ea01-416c-b893-4f824426f225, reason: Instance cf523cf4-028c-4a24-9822-df384974458e could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1351.285745] env[62952]: INFO nova.compute.manager [-] [instance: cf523cf4-028c-4a24-9822-df384974458e] Took 1.43 seconds to deallocate network for instance. [ 1351.833071] env[62952]: INFO nova.compute.manager [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: cf523cf4-028c-4a24-9822-df384974458e] Took 0.55 seconds to detach 1 volumes for instance. [ 1352.340031] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1352.340396] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1352.341029] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1352.362147] env[62952]: INFO nova.scheduler.client.report [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Deleted allocations for instance cf523cf4-028c-4a24-9822-df384974458e [ 1352.869720] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6cfc6a53-77ae-48df-a485-b97acacfec5f tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "cf523cf4-028c-4a24-9822-df384974458e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.251s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1353.654710] env[62952]: DEBUG oslo_concurrency.lockutils [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "9f3466ae-f814-4831-b485-e8c5c28e02ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1353.655018] env[62952]: DEBUG oslo_concurrency.lockutils [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "9f3466ae-f814-4831-b485-e8c5c28e02ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1354.157505] env[62952]: DEBUG nova.compute.manager [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1354.679865] env[62952]: DEBUG oslo_concurrency.lockutils [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1354.680138] env[62952]: DEBUG oslo_concurrency.lockutils [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1354.681632] env[62952]: INFO nova.compute.claims [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1355.725770] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22b6eb1f-5598-4654-b860-3974a35ada61 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.733165] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58fd274a-7753-4f56-9abc-145141d66a93 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.764238] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-903bbdf7-97b9-4430-b963-481c5e015175 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.771065] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1a60b4-8aae-4e0c-a5b0-c482fe1b06f3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.783536] env[62952]: DEBUG nova.compute.provider_tree [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1356.286426] env[62952]: DEBUG nova.scheduler.client.report [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1356.792058] env[62952]: DEBUG oslo_concurrency.lockutils [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.111s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1356.792058] env[62952]: DEBUG nova.compute.manager [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1357.296762] env[62952]: DEBUG nova.compute.utils [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1357.298237] env[62952]: DEBUG nova.compute.manager [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1357.298406] env[62952]: DEBUG nova.network.neutron [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1357.345587] env[62952]: DEBUG nova.policy [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '166cce5924454e73bc7deb1c77191f66', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e2f92c0076c14f15b5a4b8188f50db8e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1357.610630] env[62952]: DEBUG nova.network.neutron [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Successfully created port: ae6d72db-0cb4-4fc3-a905-e42445320885 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1357.801926] env[62952]: DEBUG nova.compute.manager [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1358.813018] env[62952]: DEBUG nova.compute.manager [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1358.838456] env[62952]: DEBUG nova.virt.hardware [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1358.838760] env[62952]: DEBUG nova.virt.hardware [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1358.838931] env[62952]: DEBUG nova.virt.hardware [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1358.839134] env[62952]: DEBUG nova.virt.hardware [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1358.839287] env[62952]: DEBUG nova.virt.hardware [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1358.839437] env[62952]: DEBUG nova.virt.hardware [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1358.839648] env[62952]: DEBUG nova.virt.hardware [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1358.839826] env[62952]: DEBUG nova.virt.hardware [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1358.840014] env[62952]: DEBUG nova.virt.hardware [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1358.840196] env[62952]: DEBUG nova.virt.hardware [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1358.840374] env[62952]: DEBUG nova.virt.hardware [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1358.841290] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db9154d7-018b-479c-8dba-f5e4843cfb98 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.849490] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b9382ee-e7dd-4d60-845c-938d1620db71 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.969331] env[62952]: DEBUG nova.compute.manager [req-532f0ccf-0fb3-48c0-8b69-81ce58269df8 req-6c9a7a67-46b8-48ee-a5ab-3439313f0cf6 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Received event network-vif-plugged-ae6d72db-0cb4-4fc3-a905-e42445320885 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1358.969581] env[62952]: DEBUG oslo_concurrency.lockutils [req-532f0ccf-0fb3-48c0-8b69-81ce58269df8 req-6c9a7a67-46b8-48ee-a5ab-3439313f0cf6 service nova] Acquiring lock "9f3466ae-f814-4831-b485-e8c5c28e02ea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1358.969800] env[62952]: DEBUG oslo_concurrency.lockutils [req-532f0ccf-0fb3-48c0-8b69-81ce58269df8 req-6c9a7a67-46b8-48ee-a5ab-3439313f0cf6 service nova] Lock "9f3466ae-f814-4831-b485-e8c5c28e02ea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1358.969976] env[62952]: DEBUG oslo_concurrency.lockutils [req-532f0ccf-0fb3-48c0-8b69-81ce58269df8 req-6c9a7a67-46b8-48ee-a5ab-3439313f0cf6 service nova] Lock "9f3466ae-f814-4831-b485-e8c5c28e02ea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1358.970169] env[62952]: DEBUG nova.compute.manager [req-532f0ccf-0fb3-48c0-8b69-81ce58269df8 req-6c9a7a67-46b8-48ee-a5ab-3439313f0cf6 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] No waiting events found dispatching network-vif-plugged-ae6d72db-0cb4-4fc3-a905-e42445320885 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1358.970421] env[62952]: WARNING nova.compute.manager [req-532f0ccf-0fb3-48c0-8b69-81ce58269df8 req-6c9a7a67-46b8-48ee-a5ab-3439313f0cf6 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Received unexpected event network-vif-plugged-ae6d72db-0cb4-4fc3-a905-e42445320885 for instance with vm_state building and task_state spawning. [ 1359.051050] env[62952]: DEBUG nova.network.neutron [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Successfully updated port: ae6d72db-0cb4-4fc3-a905-e42445320885 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1359.553951] env[62952]: DEBUG oslo_concurrency.lockutils [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1359.554255] env[62952]: DEBUG oslo_concurrency.lockutils [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1359.554339] env[62952]: DEBUG nova.network.neutron [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1360.085895] env[62952]: DEBUG nova.network.neutron [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1360.345750] env[62952]: DEBUG nova.network.neutron [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Updating instance_info_cache with network_info: [{"id": "ae6d72db-0cb4-4fc3-a905-e42445320885", "address": "fa:16:3e:a9:88:98", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae6d72db-0c", "ovs_interfaceid": "ae6d72db-0cb4-4fc3-a905-e42445320885", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1360.848683] env[62952]: DEBUG oslo_concurrency.lockutils [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1360.848997] env[62952]: DEBUG nova.compute.manager [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Instance network_info: |[{"id": "ae6d72db-0cb4-4fc3-a905-e42445320885", "address": "fa:16:3e:a9:88:98", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae6d72db-0c", "ovs_interfaceid": "ae6d72db-0cb4-4fc3-a905-e42445320885", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1360.849466] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:88:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3f499bc9-78da-46c1-9274-19edf26d31cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ae6d72db-0cb4-4fc3-a905-e42445320885', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1360.857034] env[62952]: DEBUG oslo.service.loopingcall [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1360.857253] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1360.857490] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0bf6faec-c421-46c7-bed8-aaed88f683b3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.878237] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1360.878237] env[62952]: value = "task-1264069" [ 1360.878237] env[62952]: _type = "Task" [ 1360.878237] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.885891] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1264069, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.002617] env[62952]: DEBUG nova.compute.manager [req-52f2d183-6c58-40a9-adfc-bbca4e9fbf9e req-987fb50e-14cb-438d-9d6d-f3a3116f8cd1 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Received event network-changed-ae6d72db-0cb4-4fc3-a905-e42445320885 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1361.002868] env[62952]: DEBUG nova.compute.manager [req-52f2d183-6c58-40a9-adfc-bbca4e9fbf9e req-987fb50e-14cb-438d-9d6d-f3a3116f8cd1 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Refreshing instance network info cache due to event network-changed-ae6d72db-0cb4-4fc3-a905-e42445320885. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1361.003346] env[62952]: DEBUG oslo_concurrency.lockutils [req-52f2d183-6c58-40a9-adfc-bbca4e9fbf9e req-987fb50e-14cb-438d-9d6d-f3a3116f8cd1 service nova] Acquiring lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1361.003527] env[62952]: DEBUG oslo_concurrency.lockutils [req-52f2d183-6c58-40a9-adfc-bbca4e9fbf9e req-987fb50e-14cb-438d-9d6d-f3a3116f8cd1 service nova] Acquired lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1361.003703] env[62952]: DEBUG nova.network.neutron [req-52f2d183-6c58-40a9-adfc-bbca4e9fbf9e req-987fb50e-14cb-438d-9d6d-f3a3116f8cd1 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Refreshing network info cache for port ae6d72db-0cb4-4fc3-a905-e42445320885 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1361.388131] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1264069, 'name': CreateVM_Task, 'duration_secs': 0.290629} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.388479] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1361.389020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1361.389201] env[62952]: DEBUG oslo_concurrency.lockutils [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1361.389521] env[62952]: DEBUG oslo_concurrency.lockutils [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1361.389773] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d1d416c-bd1a-4d7e-aebc-93858091f403 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.394216] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1361.394216] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ab719f-6507-2d21-55b2-1c33bde43ed1" [ 1361.394216] env[62952]: _type = "Task" [ 1361.394216] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1361.401448] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ab719f-6507-2d21-55b2-1c33bde43ed1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.689019] env[62952]: DEBUG nova.network.neutron [req-52f2d183-6c58-40a9-adfc-bbca4e9fbf9e req-987fb50e-14cb-438d-9d6d-f3a3116f8cd1 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Updated VIF entry in instance network info cache for port ae6d72db-0cb4-4fc3-a905-e42445320885. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1361.689385] env[62952]: DEBUG nova.network.neutron [req-52f2d183-6c58-40a9-adfc-bbca4e9fbf9e req-987fb50e-14cb-438d-9d6d-f3a3116f8cd1 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Updating instance_info_cache with network_info: [{"id": "ae6d72db-0cb4-4fc3-a905-e42445320885", "address": "fa:16:3e:a9:88:98", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae6d72db-0c", "ovs_interfaceid": "ae6d72db-0cb4-4fc3-a905-e42445320885", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1361.905142] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ab719f-6507-2d21-55b2-1c33bde43ed1, 'name': SearchDatastore_Task, 'duration_secs': 0.010408} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.905444] env[62952]: DEBUG oslo_concurrency.lockutils [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1361.905764] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1361.906030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1361.906196] env[62952]: DEBUG oslo_concurrency.lockutils [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1361.906427] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1361.906697] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-72144476-af7e-4ffd-9840-421ba501e5df {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.914298] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1361.914486] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1361.915182] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2764e2fb-554b-4996-80ba-9a08d618fb75 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.919833] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1361.919833] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5206c073-1fc7-9574-2bf7-82ad61f089ed" [ 1361.919833] env[62952]: _type = "Task" [ 1361.919833] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1361.928390] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5206c073-1fc7-9574-2bf7-82ad61f089ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1362.192587] env[62952]: DEBUG oslo_concurrency.lockutils [req-52f2d183-6c58-40a9-adfc-bbca4e9fbf9e req-987fb50e-14cb-438d-9d6d-f3a3116f8cd1 service nova] Releasing lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1362.430104] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5206c073-1fc7-9574-2bf7-82ad61f089ed, 'name': SearchDatastore_Task, 'duration_secs': 0.00794} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1362.430865] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9197ec4a-37c0-4fd1-afb1-53105b27ec26 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.435526] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1362.435526] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d83b05-6bf2-65e5-d4d0-b553dda83b6f" [ 1362.435526] env[62952]: _type = "Task" [ 1362.435526] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1362.442791] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d83b05-6bf2-65e5-d4d0-b553dda83b6f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1362.946498] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d83b05-6bf2-65e5-d4d0-b553dda83b6f, 'name': SearchDatastore_Task, 'duration_secs': 0.008811} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1362.946793] env[62952]: DEBUG oslo_concurrency.lockutils [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1362.947114] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 9f3466ae-f814-4831-b485-e8c5c28e02ea/9f3466ae-f814-4831-b485-e8c5c28e02ea.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1362.947385] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-638980c1-436d-4dae-8729-f09bdc50aeac {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.954351] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1362.954351] env[62952]: value = "task-1264070" [ 1362.954351] env[62952]: _type = "Task" [ 1362.954351] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1362.961859] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264070, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1363.464788] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264070, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.419915} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1363.465237] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore2] 9f3466ae-f814-4831-b485-e8c5c28e02ea/9f3466ae-f814-4831-b485-e8c5c28e02ea.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1363.465309] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1363.465517] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8876d41f-8d34-4e15-a034-d3669a6b9bd3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.471257] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1363.471257] env[62952]: value = "task-1264071" [ 1363.471257] env[62952]: _type = "Task" [ 1363.471257] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1363.478526] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264071, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1363.980921] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264071, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059362} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1363.980921] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1363.981276] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-271805a6-c3f7-42a0-b9e7-f352d1d72dd2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.002153] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 9f3466ae-f814-4831-b485-e8c5c28e02ea/9f3466ae-f814-4831-b485-e8c5c28e02ea.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1364.002405] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-152d7d6c-d8b7-4fd3-89fb-2e6b5702406f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.021008] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1364.021008] env[62952]: value = "task-1264072" [ 1364.021008] env[62952]: _type = "Task" [ 1364.021008] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1364.028126] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264072, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1364.531195] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264072, 'name': ReconfigVM_Task, 'duration_secs': 0.260215} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1364.531452] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 9f3466ae-f814-4831-b485-e8c5c28e02ea/9f3466ae-f814-4831-b485-e8c5c28e02ea.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1364.531915] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-def29e95-b6ad-498c-a8d1-412df54eea46 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.538558] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1364.538558] env[62952]: value = "task-1264073" [ 1364.538558] env[62952]: _type = "Task" [ 1364.538558] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1364.546406] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264073, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1365.049569] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264073, 'name': Rename_Task, 'duration_secs': 0.129196} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1365.049569] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1365.049780] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-53aeff1f-9de0-4531-9e9c-4db057e680e8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.055944] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1365.055944] env[62952]: value = "task-1264074" [ 1365.055944] env[62952]: _type = "Task" [ 1365.055944] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1365.064460] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264074, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1365.566665] env[62952]: DEBUG oslo_vmware.api [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264074, 'name': PowerOnVM_Task, 'duration_secs': 0.41816} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1365.567074] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1365.567594] env[62952]: INFO nova.compute.manager [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Took 6.75 seconds to spawn the instance on the hypervisor. [ 1365.567594] env[62952]: DEBUG nova.compute.manager [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1365.568302] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c8fe79-e89a-46d4-b994-5af2e3e6a9dd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.086256] env[62952]: INFO nova.compute.manager [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Took 11.42 seconds to build instance. [ 1366.588852] env[62952]: DEBUG oslo_concurrency.lockutils [None req-259fce30-17b1-4496-b21d-828aada85906 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "9f3466ae-f814-4831-b485-e8c5c28e02ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.934s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1366.701211] env[62952]: DEBUG nova.compute.manager [req-2e0027a7-ff8d-48d8-a8b6-8da5d14db868 req-1f7717f2-a5c0-4df0-98fa-da92ceeab565 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Received event network-changed-ae6d72db-0cb4-4fc3-a905-e42445320885 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1366.701291] env[62952]: DEBUG nova.compute.manager [req-2e0027a7-ff8d-48d8-a8b6-8da5d14db868 req-1f7717f2-a5c0-4df0-98fa-da92ceeab565 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Refreshing instance network info cache due to event network-changed-ae6d72db-0cb4-4fc3-a905-e42445320885. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1366.701478] env[62952]: DEBUG oslo_concurrency.lockutils [req-2e0027a7-ff8d-48d8-a8b6-8da5d14db868 req-1f7717f2-a5c0-4df0-98fa-da92ceeab565 service nova] Acquiring lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1366.701622] env[62952]: DEBUG oslo_concurrency.lockutils [req-2e0027a7-ff8d-48d8-a8b6-8da5d14db868 req-1f7717f2-a5c0-4df0-98fa-da92ceeab565 service nova] Acquired lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1366.701782] env[62952]: DEBUG nova.network.neutron [req-2e0027a7-ff8d-48d8-a8b6-8da5d14db868 req-1f7717f2-a5c0-4df0-98fa-da92ceeab565 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Refreshing network info cache for port ae6d72db-0cb4-4fc3-a905-e42445320885 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1367.425883] env[62952]: DEBUG nova.network.neutron [req-2e0027a7-ff8d-48d8-a8b6-8da5d14db868 req-1f7717f2-a5c0-4df0-98fa-da92ceeab565 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Updated VIF entry in instance network info cache for port ae6d72db-0cb4-4fc3-a905-e42445320885. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1367.426286] env[62952]: DEBUG nova.network.neutron [req-2e0027a7-ff8d-48d8-a8b6-8da5d14db868 req-1f7717f2-a5c0-4df0-98fa-da92ceeab565 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Updating instance_info_cache with network_info: [{"id": "ae6d72db-0cb4-4fc3-a905-e42445320885", "address": "fa:16:3e:a9:88:98", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae6d72db-0c", "ovs_interfaceid": "ae6d72db-0cb4-4fc3-a905-e42445320885", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1367.928828] env[62952]: DEBUG oslo_concurrency.lockutils [req-2e0027a7-ff8d-48d8-a8b6-8da5d14db868 req-1f7717f2-a5c0-4df0-98fa-da92ceeab565 service nova] Releasing lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1372.168460] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1372.168748] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1372.672822] env[62952]: DEBUG nova.compute.utils [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1373.175563] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1374.234416] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1374.234802] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1374.234929] env[62952]: INFO nova.compute.manager [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Attaching volume 08b9af39-6205-40f3-9a47-277b604bb801 to /dev/sdb [ 1374.266641] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-423ba6c4-636e-49c6-91c2-0d502f115171 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.273826] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4de2ba5-23be-4734-9bf7-cdc2c405064a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.286416] env[62952]: DEBUG nova.virt.block_device [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Updating existing volume attachment record: f97f188b-3d24-4976-af74-af4db75ef665 {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1378.828450] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1378.828727] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272002', 'volume_id': '08b9af39-6205-40f3-9a47-277b604bb801', 'name': 'volume-08b9af39-6205-40f3-9a47-277b604bb801', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b5da3312-a52c-4941-9c3c-247f4d93fbbf', 'attached_at': '', 'detached_at': '', 'volume_id': '08b9af39-6205-40f3-9a47-277b604bb801', 'serial': '08b9af39-6205-40f3-9a47-277b604bb801'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1378.829690] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb721c7-736f-4d54-90be-679a80ea4e6a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1378.845844] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cfa11c5-e529-4732-a2c0-fb0a943261f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1378.869148] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] volume-08b9af39-6205-40f3-9a47-277b604bb801/volume-08b9af39-6205-40f3-9a47-277b604bb801.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1378.869407] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-559eeb61-8bef-4d8b-9732-1b4df3e5b615 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1378.886843] env[62952]: DEBUG oslo_vmware.api [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1378.886843] env[62952]: value = "task-1264077" [ 1378.886843] env[62952]: _type = "Task" [ 1378.886843] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1378.894132] env[62952]: DEBUG oslo_vmware.api [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264077, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1379.396451] env[62952]: DEBUG oslo_vmware.api [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264077, 'name': ReconfigVM_Task, 'duration_secs': 0.3375} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1379.396810] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Reconfigured VM instance instance-0000006e to attach disk [datastore1] volume-08b9af39-6205-40f3-9a47-277b604bb801/volume-08b9af39-6205-40f3-9a47-277b604bb801.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1379.401373] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d56c666-0724-47d3-81ee-5cb56f09eab6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.415803] env[62952]: DEBUG oslo_vmware.api [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1379.415803] env[62952]: value = "task-1264078" [ 1379.415803] env[62952]: _type = "Task" [ 1379.415803] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1379.422916] env[62952]: DEBUG oslo_vmware.api [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264078, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1379.925246] env[62952]: DEBUG oslo_vmware.api [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264078, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1380.426962] env[62952]: DEBUG oslo_vmware.api [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264078, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1380.927943] env[62952]: DEBUG oslo_vmware.api [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264078, 'name': ReconfigVM_Task, 'duration_secs': 1.040542} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1380.928345] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272002', 'volume_id': '08b9af39-6205-40f3-9a47-277b604bb801', 'name': 'volume-08b9af39-6205-40f3-9a47-277b604bb801', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b5da3312-a52c-4941-9c3c-247f4d93fbbf', 'attached_at': '', 'detached_at': '', 'volume_id': '08b9af39-6205-40f3-9a47-277b604bb801', 'serial': '08b9af39-6205-40f3-9a47-277b604bb801'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1381.963922] env[62952]: DEBUG nova.objects.instance [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lazy-loading 'flavor' on Instance uuid b5da3312-a52c-4941-9c3c-247f4d93fbbf {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1382.469266] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6c0b22e-d442-42ea-b17a-03cee1e6dbdb tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.234s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1382.697421] env[62952]: DEBUG oslo_concurrency.lockutils [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1382.697681] env[62952]: DEBUG oslo_concurrency.lockutils [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1383.201634] env[62952]: INFO nova.compute.manager [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Detaching volume 08b9af39-6205-40f3-9a47-277b604bb801 [ 1383.230965] env[62952]: INFO nova.virt.block_device [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Attempting to driver detach volume 08b9af39-6205-40f3-9a47-277b604bb801 from mountpoint /dev/sdb [ 1383.231235] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1383.231422] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272002', 'volume_id': '08b9af39-6205-40f3-9a47-277b604bb801', 'name': 'volume-08b9af39-6205-40f3-9a47-277b604bb801', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b5da3312-a52c-4941-9c3c-247f4d93fbbf', 'attached_at': '', 'detached_at': '', 'volume_id': '08b9af39-6205-40f3-9a47-277b604bb801', 'serial': '08b9af39-6205-40f3-9a47-277b604bb801'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1383.232317] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-326d9852-74d3-46b2-99e8-500a9ef7bba5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.253305] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada19e32-aee0-4a7d-ba59-b92c0b06abe4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.259821] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f741ad8-0712-4ee9-8e20-525c83da8506 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.278879] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a231072-3e93-48a3-a37d-cf618c6f8d6a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.292427] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] The volume has not been displaced from its original location: [datastore1] volume-08b9af39-6205-40f3-9a47-277b604bb801/volume-08b9af39-6205-40f3-9a47-277b604bb801.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1383.297438] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Reconfiguring VM instance instance-0000006e to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1383.297698] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12640419-92af-4c0e-8e41-a9df625ffa53 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.314433] env[62952]: DEBUG oslo_vmware.api [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1383.314433] env[62952]: value = "task-1264079" [ 1383.314433] env[62952]: _type = "Task" [ 1383.314433] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1383.322582] env[62952]: DEBUG oslo_vmware.api [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264079, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1383.823905] env[62952]: DEBUG oslo_vmware.api [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264079, 'name': ReconfigVM_Task, 'duration_secs': 0.195126} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1383.824146] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Reconfigured VM instance instance-0000006e to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1383.828662] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bbe3dd50-60ec-4c96-b796-10c945b0977a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.843240] env[62952]: DEBUG oslo_vmware.api [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1383.843240] env[62952]: value = "task-1264080" [ 1383.843240] env[62952]: _type = "Task" [ 1383.843240] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1383.850552] env[62952]: DEBUG oslo_vmware.api [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264080, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1384.353428] env[62952]: DEBUG oslo_vmware.api [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264080, 'name': ReconfigVM_Task, 'duration_secs': 0.132883} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1384.353743] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272002', 'volume_id': '08b9af39-6205-40f3-9a47-277b604bb801', 'name': 'volume-08b9af39-6205-40f3-9a47-277b604bb801', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b5da3312-a52c-4941-9c3c-247f4d93fbbf', 'attached_at': '', 'detached_at': '', 'volume_id': '08b9af39-6205-40f3-9a47-277b604bb801', 'serial': '08b9af39-6205-40f3-9a47-277b604bb801'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1384.893696] env[62952]: DEBUG nova.objects.instance [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lazy-loading 'flavor' on Instance uuid b5da3312-a52c-4941-9c3c-247f4d93fbbf {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1385.900411] env[62952]: DEBUG oslo_concurrency.lockutils [None req-279c061f-d53d-4bad-92a5-0689875e80f6 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.202s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1386.944918] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1386.944918] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1386.945358] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1386.945358] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1386.945505] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1386.947871] env[62952]: INFO nova.compute.manager [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Terminating instance [ 1386.950156] env[62952]: DEBUG nova.compute.manager [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1386.950347] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1386.951189] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3486c389-e6fc-4f54-a41f-538915482531 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.959280] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1386.959501] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-33cf134c-6af6-432d-8ab0-008001d63b9c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.966126] env[62952]: DEBUG oslo_vmware.api [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1386.966126] env[62952]: value = "task-1264081" [ 1386.966126] env[62952]: _type = "Task" [ 1386.966126] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1386.973609] env[62952]: DEBUG oslo_vmware.api [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264081, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1387.476259] env[62952]: DEBUG oslo_vmware.api [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264081, 'name': PowerOffVM_Task, 'duration_secs': 0.161259} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1387.476553] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1387.476772] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1387.477042] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd359148-37f6-4756-a456-9953e4914fca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.891047] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1388.891427] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1388.891523] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Deleting the datastore file [datastore1] b5da3312-a52c-4941-9c3c-247f4d93fbbf {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1388.891929] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c7de37a9-ded2-469a-a018-0480abf91fa4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.898871] env[62952]: DEBUG oslo_vmware.api [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1388.898871] env[62952]: value = "task-1264083" [ 1388.898871] env[62952]: _type = "Task" [ 1388.898871] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1388.906079] env[62952]: DEBUG oslo_vmware.api [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264083, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1389.272228] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1389.408734] env[62952]: DEBUG oslo_vmware.api [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264083, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137338} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1389.408930] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1389.409138] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1389.409328] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1389.409506] env[62952]: INFO nova.compute.manager [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Took 2.46 seconds to destroy the instance on the hypervisor. [ 1389.409743] env[62952]: DEBUG oslo.service.loopingcall [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1389.409933] env[62952]: DEBUG nova.compute.manager [-] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1389.410041] env[62952]: DEBUG nova.network.neutron [-] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1389.871843] env[62952]: DEBUG nova.compute.manager [req-585bfa02-66ef-4a2d-8b3e-0d08aca73d84 req-af7beb66-60f8-449b-b97c-59db6656c2e8 service nova] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Received event network-vif-deleted-ed11c649-dae7-4ce6-95cd-5ef951865824 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1389.872097] env[62952]: INFO nova.compute.manager [req-585bfa02-66ef-4a2d-8b3e-0d08aca73d84 req-af7beb66-60f8-449b-b97c-59db6656c2e8 service nova] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Neutron deleted interface ed11c649-dae7-4ce6-95cd-5ef951865824; detaching it from the instance and deleting it from the info cache [ 1389.872290] env[62952]: DEBUG nova.network.neutron [req-585bfa02-66ef-4a2d-8b3e-0d08aca73d84 req-af7beb66-60f8-449b-b97c-59db6656c2e8 service nova] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1390.272441] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1390.272733] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1390.272733] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1390.350113] env[62952]: DEBUG nova.network.neutron [-] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1390.375066] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c7507b8-552c-4ee3-b5fb-07b09d6ad854 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.384844] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cfb374f-5d35-4763-af52-62e2b315b685 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.410366] env[62952]: DEBUG nova.compute.manager [req-585bfa02-66ef-4a2d-8b3e-0d08aca73d84 req-af7beb66-60f8-449b-b97c-59db6656c2e8 service nova] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Detach interface failed, port_id=ed11c649-dae7-4ce6-95cd-5ef951865824, reason: Instance b5da3312-a52c-4941-9c3c-247f4d93fbbf could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1390.775878] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Skipping network cache update for instance because it is being deleted. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 1390.802537] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1390.802666] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquired lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1390.802809] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Forcefully refreshing network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1390.802962] env[62952]: DEBUG nova.objects.instance [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lazy-loading 'info_cache' on Instance uuid 9f3466ae-f814-4831-b485-e8c5c28e02ea {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1390.852240] env[62952]: INFO nova.compute.manager [-] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Took 1.44 seconds to deallocate network for instance. [ 1391.358272] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1391.358678] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1391.358824] env[62952]: DEBUG nova.objects.instance [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lazy-loading 'resources' on Instance uuid b5da3312-a52c-4941-9c3c-247f4d93fbbf {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1391.907009] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb09c4b-d429-41d1-aed3-1566bb6b285d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.914423] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579eb846-dddc-4916-83e0-ce0efd63d456 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.945292] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01a2a23c-0cf0-4dfd-bcaa-967ec2178f1c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.952649] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b79a3760-4b19-453f-b681-13e1d1c6d39b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.965703] env[62952]: DEBUG nova.compute.provider_tree [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1392.468964] env[62952]: DEBUG nova.scheduler.client.report [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1392.514202] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Updating instance_info_cache with network_info: [{"id": "ae6d72db-0cb4-4fc3-a905-e42445320885", "address": "fa:16:3e:a9:88:98", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae6d72db-0c", "ovs_interfaceid": "ae6d72db-0cb4-4fc3-a905-e42445320885", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1392.973980] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.615s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1392.993670] env[62952]: INFO nova.scheduler.client.report [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Deleted allocations for instance b5da3312-a52c-4941-9c3c-247f4d93fbbf [ 1393.016171] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Releasing lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1393.016380] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Updated the network info_cache for instance {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1393.016595] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1393.016727] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1393.016874] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1393.501725] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38160581-ef73-4ac2-8d75-f6da44d95155 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "b5da3312-a52c-4941-9c3c-247f4d93fbbf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.557s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1393.519900] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1393.520125] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1393.520293] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1393.520444] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1393.521920] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42554309-8d87-48d0-bb00-75d06f05b7d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.529834] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a38034ef-ac35-4937-9df9-a5f32935d81d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.543640] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f929d16-d76a-489f-9ee1-e4123fa7f0c8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.549710] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c712c6fe-5e2f-4c82-b70b-8828e1ac85e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.578579] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181209MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1393.578729] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1393.578903] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1394.602291] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 9f3466ae-f814-4831-b485-e8c5c28e02ea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1394.602549] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1394.602623] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1394.628620] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4cdc3f-f51d-47be-959f-178cc96a27ee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.635658] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1778f4dd-7404-4d8e-b5a8-79092325092d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.664436] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62446e3d-4ead-4023-8287-18827fe90a58 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.671357] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06dcc521-facb-4ef3-bb50-87cb3168811a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.684819] env[62952]: DEBUG nova.compute.provider_tree [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1395.187721] env[62952]: DEBUG nova.scheduler.client.report [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1395.693204] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1395.693517] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.115s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1395.949354] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1396.115054] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1396.115292] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1396.454633] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1396.454823] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1396.454981] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1396.618512] env[62952]: DEBUG nova.compute.manager [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1397.141672] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1397.141941] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1397.143469] env[62952]: INFO nova.compute.claims [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1398.189868] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98f65f52-1e79-4ff8-b9c5-84408ce81bdb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.197453] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7580297c-309c-4752-933c-e1aa8c4cac00 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.227176] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81de9250-87e6-47d1-9c00-a862fb73f489 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.233947] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b095e8-3c9b-4470-8a33-cc5148de2b2d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.246496] env[62952]: DEBUG nova.compute.provider_tree [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1398.271619] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1398.750382] env[62952]: DEBUG nova.scheduler.client.report [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1399.255584] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.113s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1399.256145] env[62952]: DEBUG nova.compute.manager [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1399.267536] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1399.760938] env[62952]: DEBUG nova.compute.utils [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1399.762398] env[62952]: DEBUG nova.compute.manager [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1399.762648] env[62952]: DEBUG nova.network.neutron [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1399.813152] env[62952]: DEBUG nova.policy [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0a36328ec404c22adf6d2d256c9514c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f725be1feb8a44478d297b0be58c59ba', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:203}} [ 1400.090846] env[62952]: DEBUG nova.network.neutron [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Successfully created port: 6c006fd9-5eb1-4220-9160-af6ec0ca7156 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1400.266298] env[62952]: DEBUG nova.compute.manager [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1401.275668] env[62952]: DEBUG nova.compute.manager [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1401.301751] env[62952]: DEBUG nova.virt.hardware [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T21:28:48Z,direct_url=,disk_format='vmdk',id=119dd2b4-b6de-42d5-adba-db818e4c2cd7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73931ed9ad374fb0870629249c5f6909',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T21:28:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1401.302066] env[62952]: DEBUG nova.virt.hardware [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1401.302239] env[62952]: DEBUG nova.virt.hardware [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1401.302423] env[62952]: DEBUG nova.virt.hardware [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1401.302572] env[62952]: DEBUG nova.virt.hardware [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1401.302724] env[62952]: DEBUG nova.virt.hardware [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1401.302936] env[62952]: DEBUG nova.virt.hardware [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1401.303121] env[62952]: DEBUG nova.virt.hardware [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1401.303298] env[62952]: DEBUG nova.virt.hardware [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1401.303463] env[62952]: DEBUG nova.virt.hardware [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1401.303635] env[62952]: DEBUG nova.virt.hardware [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1401.304500] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f03f8038-9333-4071-bbb3-ca11e4e5a600 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.312574] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0126aac0-5924-49fd-88c1-9030ec582c26 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.431207] env[62952]: DEBUG nova.compute.manager [req-c0345940-be66-4031-9ee8-b5fa35425cce req-623756c0-7fbb-4f36-9c2e-78863617d346 service nova] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Received event network-vif-plugged-6c006fd9-5eb1-4220-9160-af6ec0ca7156 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1401.431207] env[62952]: DEBUG oslo_concurrency.lockutils [req-c0345940-be66-4031-9ee8-b5fa35425cce req-623756c0-7fbb-4f36-9c2e-78863617d346 service nova] Acquiring lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1401.431207] env[62952]: DEBUG oslo_concurrency.lockutils [req-c0345940-be66-4031-9ee8-b5fa35425cce req-623756c0-7fbb-4f36-9c2e-78863617d346 service nova] Lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1401.431207] env[62952]: DEBUG oslo_concurrency.lockutils [req-c0345940-be66-4031-9ee8-b5fa35425cce req-623756c0-7fbb-4f36-9c2e-78863617d346 service nova] Lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1401.431207] env[62952]: DEBUG nova.compute.manager [req-c0345940-be66-4031-9ee8-b5fa35425cce req-623756c0-7fbb-4f36-9c2e-78863617d346 service nova] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] No waiting events found dispatching network-vif-plugged-6c006fd9-5eb1-4220-9160-af6ec0ca7156 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1401.431207] env[62952]: WARNING nova.compute.manager [req-c0345940-be66-4031-9ee8-b5fa35425cce req-623756c0-7fbb-4f36-9c2e-78863617d346 service nova] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Received unexpected event network-vif-plugged-6c006fd9-5eb1-4220-9160-af6ec0ca7156 for instance with vm_state building and task_state spawning. [ 1401.984459] env[62952]: DEBUG nova.network.neutron [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Successfully updated port: 6c006fd9-5eb1-4220-9160-af6ec0ca7156 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1402.007445] env[62952]: DEBUG nova.compute.manager [req-01c4a7e6-d737-4a83-8f52-9090edc5ef1c req-2aab9379-63e9-4142-808e-e67e33499f76 service nova] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Received event network-changed-6c006fd9-5eb1-4220-9160-af6ec0ca7156 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1402.007656] env[62952]: DEBUG nova.compute.manager [req-01c4a7e6-d737-4a83-8f52-9090edc5ef1c req-2aab9379-63e9-4142-808e-e67e33499f76 service nova] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Refreshing instance network info cache due to event network-changed-6c006fd9-5eb1-4220-9160-af6ec0ca7156. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1402.007902] env[62952]: DEBUG oslo_concurrency.lockutils [req-01c4a7e6-d737-4a83-8f52-9090edc5ef1c req-2aab9379-63e9-4142-808e-e67e33499f76 service nova] Acquiring lock "refresh_cache-7bf9931c-18a3-4bb5-99d2-a45263c6b236" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1402.008034] env[62952]: DEBUG oslo_concurrency.lockutils [req-01c4a7e6-d737-4a83-8f52-9090edc5ef1c req-2aab9379-63e9-4142-808e-e67e33499f76 service nova] Acquired lock "refresh_cache-7bf9931c-18a3-4bb5-99d2-a45263c6b236" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1402.008174] env[62952]: DEBUG nova.network.neutron [req-01c4a7e6-d737-4a83-8f52-9090edc5ef1c req-2aab9379-63e9-4142-808e-e67e33499f76 service nova] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Refreshing network info cache for port 6c006fd9-5eb1-4220-9160-af6ec0ca7156 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1402.486093] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "refresh_cache-7bf9931c-18a3-4bb5-99d2-a45263c6b236" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1402.542877] env[62952]: DEBUG nova.network.neutron [req-01c4a7e6-d737-4a83-8f52-9090edc5ef1c req-2aab9379-63e9-4142-808e-e67e33499f76 service nova] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1402.622190] env[62952]: DEBUG nova.network.neutron [req-01c4a7e6-d737-4a83-8f52-9090edc5ef1c req-2aab9379-63e9-4142-808e-e67e33499f76 service nova] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1403.125417] env[62952]: DEBUG oslo_concurrency.lockutils [req-01c4a7e6-d737-4a83-8f52-9090edc5ef1c req-2aab9379-63e9-4142-808e-e67e33499f76 service nova] Releasing lock "refresh_cache-7bf9931c-18a3-4bb5-99d2-a45263c6b236" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1403.125823] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquired lock "refresh_cache-7bf9931c-18a3-4bb5-99d2-a45263c6b236" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1403.125983] env[62952]: DEBUG nova.network.neutron [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1403.390242] env[62952]: DEBUG oslo_concurrency.lockutils [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "9f3466ae-f814-4831-b485-e8c5c28e02ea" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1403.390507] env[62952]: DEBUG oslo_concurrency.lockutils [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "9f3466ae-f814-4831-b485-e8c5c28e02ea" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1403.390638] env[62952]: INFO nova.compute.manager [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Shelving [ 1403.657088] env[62952]: DEBUG nova.network.neutron [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1403.786524] env[62952]: DEBUG nova.network.neutron [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Updating instance_info_cache with network_info: [{"id": "6c006fd9-5eb1-4220-9160-af6ec0ca7156", "address": "fa:16:3e:3c:9b:34", "network": {"id": "8e167a8b-2628-43e1-8418-e673064745b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-232555460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f725be1feb8a44478d297b0be58c59ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c006fd9-5e", "ovs_interfaceid": "6c006fd9-5eb1-4220-9160-af6ec0ca7156", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1403.898763] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1403.899029] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2faba031-c3f3-41b1-8f1b-b6053e46adf0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.905674] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1403.905674] env[62952]: value = "task-1264084" [ 1403.905674] env[62952]: _type = "Task" [ 1403.905674] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1403.913240] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264084, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1404.289691] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Releasing lock "refresh_cache-7bf9931c-18a3-4bb5-99d2-a45263c6b236" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1404.290061] env[62952]: DEBUG nova.compute.manager [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Instance network_info: |[{"id": "6c006fd9-5eb1-4220-9160-af6ec0ca7156", "address": "fa:16:3e:3c:9b:34", "network": {"id": "8e167a8b-2628-43e1-8418-e673064745b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-232555460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f725be1feb8a44478d297b0be58c59ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c006fd9-5e", "ovs_interfaceid": "6c006fd9-5eb1-4220-9160-af6ec0ca7156", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1404.290509] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:9b:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69351262-8d39-441a-85ba-3a78df436d17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6c006fd9-5eb1-4220-9160-af6ec0ca7156', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1404.298057] env[62952]: DEBUG oslo.service.loopingcall [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1404.298275] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1404.298499] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d729c0a5-2386-451d-a851-c5d26e7b7611 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.318123] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1404.318123] env[62952]: value = "task-1264085" [ 1404.318123] env[62952]: _type = "Task" [ 1404.318123] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1404.325267] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1264085, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1404.417066] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264084, 'name': PowerOffVM_Task, 'duration_secs': 0.168734} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1404.417354] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1404.418130] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5038e863-d3a3-419e-8787-0465345764d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.439923] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-959a49c7-dff8-4297-97f1-95a5291086bc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.827926] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1264085, 'name': CreateVM_Task, 'duration_secs': 0.291443} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1404.828369] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1404.828784] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1404.828958] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1404.829303] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1404.829544] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdc2f67c-0da3-4a8b-8d17-ae686c688395 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.833642] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1404.833642] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d387fa-706c-e371-af06-c7d6b03b2c4f" [ 1404.833642] env[62952]: _type = "Task" [ 1404.833642] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1404.840709] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d387fa-706c-e371-af06-c7d6b03b2c4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1404.951755] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1404.952078] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9e8381e3-49b6-4c5a-a9b5-ddf834b5f0ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.959136] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1404.959136] env[62952]: value = "task-1264086" [ 1404.959136] env[62952]: _type = "Task" [ 1404.959136] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1404.966995] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264086, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.346214] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52d387fa-706c-e371-af06-c7d6b03b2c4f, 'name': SearchDatastore_Task, 'duration_secs': 0.009464} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1405.346619] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1405.346942] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Processing image 119dd2b4-b6de-42d5-adba-db818e4c2cd7 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1405.347251] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1405.347451] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquired lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1405.347682] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1405.347993] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-993d151b-1552-40da-8a03-e5089a847d37 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.355952] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1405.356192] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1405.356962] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-359443d5-414a-4a27-b24d-fbfa2e8a6bb5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.361907] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1405.361907] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5216dce8-969e-d982-f65a-8a0d3a6b67e7" [ 1405.361907] env[62952]: _type = "Task" [ 1405.361907] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1405.369975] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5216dce8-969e-d982-f65a-8a0d3a6b67e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.469175] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264086, 'name': CreateSnapshot_Task, 'duration_secs': 0.406765} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1405.469455] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1405.470224] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc18676-6249-4fb5-95b3-82c566c41473 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.871999] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]5216dce8-969e-d982-f65a-8a0d3a6b67e7, 'name': SearchDatastore_Task, 'duration_secs': 0.008368} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1405.872768] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f26299ac-db8b-43ba-b480-633525f86c9b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.877434] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1405.877434] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52580ab1-07e2-6e9b-52b9-ad77f9c73768" [ 1405.877434] env[62952]: _type = "Task" [ 1405.877434] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1405.884474] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52580ab1-07e2-6e9b-52b9-ad77f9c73768, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.987757] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1405.988022] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-721550c3-5950-4252-a5f1-39fbe2c9bbfc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.995935] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1405.995935] env[62952]: value = "task-1264087" [ 1405.995935] env[62952]: _type = "Task" [ 1405.995935] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1406.004988] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264087, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1406.388330] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52580ab1-07e2-6e9b-52b9-ad77f9c73768, 'name': SearchDatastore_Task, 'duration_secs': 0.008759} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1406.388621] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Releasing lock "[datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1406.388912] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 7bf9931c-18a3-4bb5-99d2-a45263c6b236/7bf9931c-18a3-4bb5-99d2-a45263c6b236.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1406.389218] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3246150f-1116-4879-85a0-42cd99e4c0ae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.395205] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1406.395205] env[62952]: value = "task-1264088" [ 1406.395205] env[62952]: _type = "Task" [ 1406.395205] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1406.403045] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264088, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1406.505811] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264087, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1406.904798] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264088, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455351} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1406.905950] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/119dd2b4-b6de-42d5-adba-db818e4c2cd7/119dd2b4-b6de-42d5-adba-db818e4c2cd7.vmdk to [datastore1] 7bf9931c-18a3-4bb5-99d2-a45263c6b236/7bf9931c-18a3-4bb5-99d2-a45263c6b236.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1406.905950] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1406.906196] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0904a863-ad6d-4441-aacc-414330386552 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.912274] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1406.912274] env[62952]: value = "task-1264089" [ 1406.912274] env[62952]: _type = "Task" [ 1406.912274] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1406.921571] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264089, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1407.005705] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264087, 'name': CloneVM_Task, 'duration_secs': 0.946718} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1407.005965] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Created linked-clone VM from snapshot [ 1407.006702] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283f575e-100e-4707-b640-bbe4f25d39f5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.013357] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Uploading image 5686e478-a71e-4391-b5e5-6826db862980 {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1407.036562] env[62952]: DEBUG oslo_vmware.rw_handles [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1407.036562] env[62952]: value = "vm-272005" [ 1407.036562] env[62952]: _type = "VirtualMachine" [ 1407.036562] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1407.036812] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c911e55a-a88e-4b34-9577-a91174df7ffa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.043296] env[62952]: DEBUG oslo_vmware.rw_handles [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lease: (returnval){ [ 1407.043296] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a5585d-1e70-99c1-5998-41d57f0d62e2" [ 1407.043296] env[62952]: _type = "HttpNfcLease" [ 1407.043296] env[62952]: } obtained for exporting VM: (result){ [ 1407.043296] env[62952]: value = "vm-272005" [ 1407.043296] env[62952]: _type = "VirtualMachine" [ 1407.043296] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1407.043568] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the lease: (returnval){ [ 1407.043568] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a5585d-1e70-99c1-5998-41d57f0d62e2" [ 1407.043568] env[62952]: _type = "HttpNfcLease" [ 1407.043568] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1407.049093] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1407.049093] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a5585d-1e70-99c1-5998-41d57f0d62e2" [ 1407.049093] env[62952]: _type = "HttpNfcLease" [ 1407.049093] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1407.422650] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264089, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079388} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1407.422938] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1407.423724] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc7e752-e433-4fb3-be75-1f40994f9430 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.444668] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] 7bf9931c-18a3-4bb5-99d2-a45263c6b236/7bf9931c-18a3-4bb5-99d2-a45263c6b236.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1407.444909] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3da52e54-fe0e-4b5e-9e91-e99f872069c1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.463612] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1407.463612] env[62952]: value = "task-1264091" [ 1407.463612] env[62952]: _type = "Task" [ 1407.463612] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1407.470945] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264091, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1407.552055] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1407.552055] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a5585d-1e70-99c1-5998-41d57f0d62e2" [ 1407.552055] env[62952]: _type = "HttpNfcLease" [ 1407.552055] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1407.552459] env[62952]: DEBUG oslo_vmware.rw_handles [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1407.552459] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52a5585d-1e70-99c1-5998-41d57f0d62e2" [ 1407.552459] env[62952]: _type = "HttpNfcLease" [ 1407.552459] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1407.553278] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2180bee1-8a4b-455f-8ae2-3425c46decf8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.562597] env[62952]: DEBUG oslo_vmware.rw_handles [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524f3973-a28b-36fe-e393-90b8075e465a/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1407.562889] env[62952]: DEBUG oslo_vmware.rw_handles [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524f3973-a28b-36fe-e393-90b8075e465a/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1407.651023] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-cbd9b6e7-4519-400d-8760-453f9799e668 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.973988] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264091, 'name': ReconfigVM_Task, 'duration_secs': 0.319107} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1407.974428] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Reconfigured VM instance instance-00000070 to attach disk [datastore1] 7bf9931c-18a3-4bb5-99d2-a45263c6b236/7bf9931c-18a3-4bb5-99d2-a45263c6b236.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1407.975346] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e8dbdb02-f0e5-492c-a430-fdea3c57cfc0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.982178] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1407.982178] env[62952]: value = "task-1264092" [ 1407.982178] env[62952]: _type = "Task" [ 1407.982178] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1407.992585] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264092, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1408.492847] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264092, 'name': Rename_Task, 'duration_secs': 0.131254} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1408.493256] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1408.493571] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f076ab5c-126c-42d9-83df-4db4325483c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.500429] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1408.500429] env[62952]: value = "task-1264093" [ 1408.500429] env[62952]: _type = "Task" [ 1408.500429] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1408.508376] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264093, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1409.010934] env[62952]: DEBUG oslo_vmware.api [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264093, 'name': PowerOnVM_Task, 'duration_secs': 0.484328} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1409.011970] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1409.012204] env[62952]: INFO nova.compute.manager [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Took 7.74 seconds to spawn the instance on the hypervisor. [ 1409.012510] env[62952]: DEBUG nova.compute.manager [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1409.013356] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-181c5ab9-1714-4ad3-8d68-0bb5f1755b79 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.534092] env[62952]: INFO nova.compute.manager [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Took 12.41 seconds to build instance. [ 1410.037242] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8eb2de12-7247-4efe-aca3-1fba6d885fb9 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.921s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1410.328055] env[62952]: DEBUG nova.compute.manager [req-3c56fd88-dfc4-4d73-ae23-c78ce8605e90 req-1c64637a-a32f-4b12-a5b7-3e7635c0a71c service nova] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Received event network-changed-6c006fd9-5eb1-4220-9160-af6ec0ca7156 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1410.328464] env[62952]: DEBUG nova.compute.manager [req-3c56fd88-dfc4-4d73-ae23-c78ce8605e90 req-1c64637a-a32f-4b12-a5b7-3e7635c0a71c service nova] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Refreshing instance network info cache due to event network-changed-6c006fd9-5eb1-4220-9160-af6ec0ca7156. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1410.328850] env[62952]: DEBUG oslo_concurrency.lockutils [req-3c56fd88-dfc4-4d73-ae23-c78ce8605e90 req-1c64637a-a32f-4b12-a5b7-3e7635c0a71c service nova] Acquiring lock "refresh_cache-7bf9931c-18a3-4bb5-99d2-a45263c6b236" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1410.329141] env[62952]: DEBUG oslo_concurrency.lockutils [req-3c56fd88-dfc4-4d73-ae23-c78ce8605e90 req-1c64637a-a32f-4b12-a5b7-3e7635c0a71c service nova] Acquired lock "refresh_cache-7bf9931c-18a3-4bb5-99d2-a45263c6b236" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1410.329443] env[62952]: DEBUG nova.network.neutron [req-3c56fd88-dfc4-4d73-ae23-c78ce8605e90 req-1c64637a-a32f-4b12-a5b7-3e7635c0a71c service nova] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Refreshing network info cache for port 6c006fd9-5eb1-4220-9160-af6ec0ca7156 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1411.060321] env[62952]: DEBUG nova.network.neutron [req-3c56fd88-dfc4-4d73-ae23-c78ce8605e90 req-1c64637a-a32f-4b12-a5b7-3e7635c0a71c service nova] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Updated VIF entry in instance network info cache for port 6c006fd9-5eb1-4220-9160-af6ec0ca7156. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1411.060741] env[62952]: DEBUG nova.network.neutron [req-3c56fd88-dfc4-4d73-ae23-c78ce8605e90 req-1c64637a-a32f-4b12-a5b7-3e7635c0a71c service nova] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Updating instance_info_cache with network_info: [{"id": "6c006fd9-5eb1-4220-9160-af6ec0ca7156", "address": "fa:16:3e:3c:9b:34", "network": {"id": "8e167a8b-2628-43e1-8418-e673064745b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-232555460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f725be1feb8a44478d297b0be58c59ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c006fd9-5e", "ovs_interfaceid": "6c006fd9-5eb1-4220-9160-af6ec0ca7156", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1411.563769] env[62952]: DEBUG oslo_concurrency.lockutils [req-3c56fd88-dfc4-4d73-ae23-c78ce8605e90 req-1c64637a-a32f-4b12-a5b7-3e7635c0a71c service nova] Releasing lock "refresh_cache-7bf9931c-18a3-4bb5-99d2-a45263c6b236" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1415.179912] env[62952]: DEBUG oslo_vmware.rw_handles [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524f3973-a28b-36fe-e393-90b8075e465a/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1415.180892] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d87624-9c7e-4a21-b44a-60eb2da91f21 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.187422] env[62952]: DEBUG oslo_vmware.rw_handles [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524f3973-a28b-36fe-e393-90b8075e465a/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1415.187599] env[62952]: ERROR oslo_vmware.rw_handles [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524f3973-a28b-36fe-e393-90b8075e465a/disk-0.vmdk due to incomplete transfer. [ 1415.187817] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-2f443df0-f055-4585-a13d-3fd0a988f4d1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.195207] env[62952]: DEBUG oslo_vmware.rw_handles [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524f3973-a28b-36fe-e393-90b8075e465a/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1415.195400] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Uploaded image 5686e478-a71e-4391-b5e5-6826db862980 to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1415.197631] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1415.197859] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9afdaea5-6f92-487d-a4b5-8f32171a7a6b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.203946] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1415.203946] env[62952]: value = "task-1264094" [ 1415.203946] env[62952]: _type = "Task" [ 1415.203946] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1415.211682] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264094, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1415.713967] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264094, 'name': Destroy_Task, 'duration_secs': 0.304823} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1415.714247] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Destroyed the VM [ 1415.714488] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1415.714731] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-337859d1-5538-4d14-8a21-14c03f0a3c36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.721202] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1415.721202] env[62952]: value = "task-1264095" [ 1415.721202] env[62952]: _type = "Task" [ 1415.721202] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1415.728277] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264095, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1416.230929] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264095, 'name': RemoveSnapshot_Task, 'duration_secs': 0.309846} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1416.231325] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1416.231530] env[62952]: DEBUG nova.compute.manager [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1416.232286] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a752b9-e4d9-4a4a-9641-8f9af79b3ad2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.744372] env[62952]: INFO nova.compute.manager [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Shelve offloading [ 1416.746313] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1416.746606] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47f834dd-4134-46e4-88ea-675d12e9bd90 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.754019] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1416.754019] env[62952]: value = "task-1264096" [ 1416.754019] env[62952]: _type = "Task" [ 1416.754019] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1416.762359] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264096, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1417.265044] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1417.265491] env[62952]: DEBUG nova.compute.manager [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1417.265981] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be7742f6-1437-4c88-8a05-ac7f57980a90 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.271204] env[62952]: DEBUG oslo_concurrency.lockutils [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1417.271369] env[62952]: DEBUG oslo_concurrency.lockutils [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1417.271539] env[62952]: DEBUG nova.network.neutron [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1417.985345] env[62952]: DEBUG nova.network.neutron [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Updating instance_info_cache with network_info: [{"id": "ae6d72db-0cb4-4fc3-a905-e42445320885", "address": "fa:16:3e:a9:88:98", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae6d72db-0c", "ovs_interfaceid": "ae6d72db-0cb4-4fc3-a905-e42445320885", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1418.487765] env[62952]: DEBUG oslo_concurrency.lockutils [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1418.719783] env[62952]: DEBUG nova.compute.manager [req-cfc46d36-76f9-4c64-8809-abb7bafdc96f req-e556e5e8-88dc-4fe4-afaa-6a946881e060 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Received event network-vif-unplugged-ae6d72db-0cb4-4fc3-a905-e42445320885 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1418.720042] env[62952]: DEBUG oslo_concurrency.lockutils [req-cfc46d36-76f9-4c64-8809-abb7bafdc96f req-e556e5e8-88dc-4fe4-afaa-6a946881e060 service nova] Acquiring lock "9f3466ae-f814-4831-b485-e8c5c28e02ea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1418.720270] env[62952]: DEBUG oslo_concurrency.lockutils [req-cfc46d36-76f9-4c64-8809-abb7bafdc96f req-e556e5e8-88dc-4fe4-afaa-6a946881e060 service nova] Lock "9f3466ae-f814-4831-b485-e8c5c28e02ea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1418.720446] env[62952]: DEBUG oslo_concurrency.lockutils [req-cfc46d36-76f9-4c64-8809-abb7bafdc96f req-e556e5e8-88dc-4fe4-afaa-6a946881e060 service nova] Lock "9f3466ae-f814-4831-b485-e8c5c28e02ea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1418.720617] env[62952]: DEBUG nova.compute.manager [req-cfc46d36-76f9-4c64-8809-abb7bafdc96f req-e556e5e8-88dc-4fe4-afaa-6a946881e060 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] No waiting events found dispatching network-vif-unplugged-ae6d72db-0cb4-4fc3-a905-e42445320885 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1418.720791] env[62952]: WARNING nova.compute.manager [req-cfc46d36-76f9-4c64-8809-abb7bafdc96f req-e556e5e8-88dc-4fe4-afaa-6a946881e060 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Received unexpected event network-vif-unplugged-ae6d72db-0cb4-4fc3-a905-e42445320885 for instance with vm_state shelved and task_state shelving_offloading. [ 1418.804240] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1418.805150] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56bba82a-dbcd-4737-9d01-947d14c7de4e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.812560] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1418.812777] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-83a1ef0a-4bd2-4a94-b787-a5261ff740ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.878657] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1418.878870] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1418.879108] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Deleting the datastore file [datastore2] 9f3466ae-f814-4831-b485-e8c5c28e02ea {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1418.879370] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d576f9ce-af56-4701-80ac-bbe94223f875 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.885527] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1418.885527] env[62952]: value = "task-1264098" [ 1418.885527] env[62952]: _type = "Task" [ 1418.885527] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1418.893095] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264098, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1419.396103] env[62952]: DEBUG oslo_vmware.api [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264098, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122837} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1419.396384] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1419.396567] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1419.396738] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1419.421089] env[62952]: INFO nova.scheduler.client.report [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Deleted allocations for instance 9f3466ae-f814-4831-b485-e8c5c28e02ea [ 1419.925583] env[62952]: DEBUG oslo_concurrency.lockutils [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1419.925968] env[62952]: DEBUG oslo_concurrency.lockutils [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1419.926173] env[62952]: DEBUG nova.objects.instance [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lazy-loading 'resources' on Instance uuid 9f3466ae-f814-4831-b485-e8c5c28e02ea {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1420.429304] env[62952]: DEBUG nova.objects.instance [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lazy-loading 'numa_topology' on Instance uuid 9f3466ae-f814-4831-b485-e8c5c28e02ea {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1420.747296] env[62952]: DEBUG nova.compute.manager [req-6b03c133-46b7-45f2-9dab-e25f618624a1 req-a654ab91-7696-4df9-b194-fdd28eec3b27 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Received event network-changed-ae6d72db-0cb4-4fc3-a905-e42445320885 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1420.747504] env[62952]: DEBUG nova.compute.manager [req-6b03c133-46b7-45f2-9dab-e25f618624a1 req-a654ab91-7696-4df9-b194-fdd28eec3b27 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Refreshing instance network info cache due to event network-changed-ae6d72db-0cb4-4fc3-a905-e42445320885. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1420.747720] env[62952]: DEBUG oslo_concurrency.lockutils [req-6b03c133-46b7-45f2-9dab-e25f618624a1 req-a654ab91-7696-4df9-b194-fdd28eec3b27 service nova] Acquiring lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1420.747863] env[62952]: DEBUG oslo_concurrency.lockutils [req-6b03c133-46b7-45f2-9dab-e25f618624a1 req-a654ab91-7696-4df9-b194-fdd28eec3b27 service nova] Acquired lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1420.748033] env[62952]: DEBUG nova.network.neutron [req-6b03c133-46b7-45f2-9dab-e25f618624a1 req-a654ab91-7696-4df9-b194-fdd28eec3b27 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Refreshing network info cache for port ae6d72db-0cb4-4fc3-a905-e42445320885 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1420.932045] env[62952]: DEBUG nova.objects.base [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Object Instance<9f3466ae-f814-4831-b485-e8c5c28e02ea> lazy-loaded attributes: resources,numa_topology {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1420.969812] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc227b5-4e95-47ff-a02b-51d45c2207f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.977600] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d59cf92-4afc-4c1a-abbb-6be0bd7ec2cd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.006477] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f431d4a-d845-4974-903a-f5c0c28e4791 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.013320] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a332eb56-d716-4e0b-87d4-1a04af2bed94 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.026257] env[62952]: DEBUG nova.compute.provider_tree [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1421.179025] env[62952]: DEBUG oslo_concurrency.lockutils [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "9f3466ae-f814-4831-b485-e8c5c28e02ea" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1421.457982] env[62952]: DEBUG nova.network.neutron [req-6b03c133-46b7-45f2-9dab-e25f618624a1 req-a654ab91-7696-4df9-b194-fdd28eec3b27 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Updated VIF entry in instance network info cache for port ae6d72db-0cb4-4fc3-a905-e42445320885. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1421.458391] env[62952]: DEBUG nova.network.neutron [req-6b03c133-46b7-45f2-9dab-e25f618624a1 req-a654ab91-7696-4df9-b194-fdd28eec3b27 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Updating instance_info_cache with network_info: [{"id": "ae6d72db-0cb4-4fc3-a905-e42445320885", "address": "fa:16:3e:a9:88:98", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapae6d72db-0c", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1421.529470] env[62952]: DEBUG nova.scheduler.client.report [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1421.961778] env[62952]: DEBUG oslo_concurrency.lockutils [req-6b03c133-46b7-45f2-9dab-e25f618624a1 req-a654ab91-7696-4df9-b194-fdd28eec3b27 service nova] Releasing lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1422.033927] env[62952]: DEBUG oslo_concurrency.lockutils [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.108s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1422.542231] env[62952]: DEBUG oslo_concurrency.lockutils [None req-889ca9a9-7de9-4cef-9a99-800d679d0326 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "9f3466ae-f814-4831-b485-e8c5c28e02ea" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.152s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1422.543138] env[62952]: DEBUG oslo_concurrency.lockutils [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "9f3466ae-f814-4831-b485-e8c5c28e02ea" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.364s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1422.543383] env[62952]: INFO nova.compute.manager [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Unshelving [ 1423.568093] env[62952]: DEBUG oslo_concurrency.lockutils [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1423.568420] env[62952]: DEBUG oslo_concurrency.lockutils [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1423.568586] env[62952]: DEBUG nova.objects.instance [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lazy-loading 'pci_requests' on Instance uuid 9f3466ae-f814-4831-b485-e8c5c28e02ea {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1424.072735] env[62952]: DEBUG nova.objects.instance [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lazy-loading 'numa_topology' on Instance uuid 9f3466ae-f814-4831-b485-e8c5c28e02ea {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1424.575129] env[62952]: INFO nova.compute.claims [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1425.620178] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c30e83-059e-4675-9038-b3298534036c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.627880] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a67f211-5089-47c4-ac68-5c727868782d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.657073] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8803ad04-aab8-4a4d-96f4-cc9c3c6894de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.663578] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d5b22bb-cab1-4f58-a52a-98586f8f4a3b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.675869] env[62952]: DEBUG nova.compute.provider_tree [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1426.178764] env[62952]: DEBUG nova.scheduler.client.report [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1426.683561] env[62952]: DEBUG oslo_concurrency.lockutils [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.115s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1426.711517] env[62952]: INFO nova.network.neutron [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Updating port ae6d72db-0cb4-4fc3-a905-e42445320885 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1428.079158] env[62952]: DEBUG nova.compute.manager [req-7d6b5d35-25df-491e-80f3-97c07fd1cc11 req-21ae3962-b926-467c-969a-4cb42bdbda27 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Received event network-vif-plugged-ae6d72db-0cb4-4fc3-a905-e42445320885 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1428.079429] env[62952]: DEBUG oslo_concurrency.lockutils [req-7d6b5d35-25df-491e-80f3-97c07fd1cc11 req-21ae3962-b926-467c-969a-4cb42bdbda27 service nova] Acquiring lock "9f3466ae-f814-4831-b485-e8c5c28e02ea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1428.079573] env[62952]: DEBUG oslo_concurrency.lockutils [req-7d6b5d35-25df-491e-80f3-97c07fd1cc11 req-21ae3962-b926-467c-969a-4cb42bdbda27 service nova] Lock "9f3466ae-f814-4831-b485-e8c5c28e02ea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1428.079744] env[62952]: DEBUG oslo_concurrency.lockutils [req-7d6b5d35-25df-491e-80f3-97c07fd1cc11 req-21ae3962-b926-467c-969a-4cb42bdbda27 service nova] Lock "9f3466ae-f814-4831-b485-e8c5c28e02ea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1428.079915] env[62952]: DEBUG nova.compute.manager [req-7d6b5d35-25df-491e-80f3-97c07fd1cc11 req-21ae3962-b926-467c-969a-4cb42bdbda27 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] No waiting events found dispatching network-vif-plugged-ae6d72db-0cb4-4fc3-a905-e42445320885 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1428.080097] env[62952]: WARNING nova.compute.manager [req-7d6b5d35-25df-491e-80f3-97c07fd1cc11 req-21ae3962-b926-467c-969a-4cb42bdbda27 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Received unexpected event network-vif-plugged-ae6d72db-0cb4-4fc3-a905-e42445320885 for instance with vm_state shelved_offloaded and task_state spawning. [ 1428.171704] env[62952]: DEBUG oslo_concurrency.lockutils [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1428.171704] env[62952]: DEBUG oslo_concurrency.lockutils [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1428.171990] env[62952]: DEBUG nova.network.neutron [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1428.869042] env[62952]: DEBUG nova.network.neutron [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Updating instance_info_cache with network_info: [{"id": "ae6d72db-0cb4-4fc3-a905-e42445320885", "address": "fa:16:3e:a9:88:98", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae6d72db-0c", "ovs_interfaceid": "ae6d72db-0cb4-4fc3-a905-e42445320885", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1429.371774] env[62952]: DEBUG oslo_concurrency.lockutils [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1429.397649] env[62952]: DEBUG nova.virt.hardware [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T21:29:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='8adb76d7f767810d7af6eccd3d746516',container_format='bare',created_at=2024-10-10T21:44:15Z,direct_url=,disk_format='vmdk',id=5686e478-a71e-4391-b5e5-6826db862980,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-2099351937-shelved',owner='e2f92c0076c14f15b5a4b8188f50db8e',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-10-10T21:44:27Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1429.397885] env[62952]: DEBUG nova.virt.hardware [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1429.398057] env[62952]: DEBUG nova.virt.hardware [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1429.398253] env[62952]: DEBUG nova.virt.hardware [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1429.398403] env[62952]: DEBUG nova.virt.hardware [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1429.398551] env[62952]: DEBUG nova.virt.hardware [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1429.398758] env[62952]: DEBUG nova.virt.hardware [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1429.398917] env[62952]: DEBUG nova.virt.hardware [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1429.399097] env[62952]: DEBUG nova.virt.hardware [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1429.399299] env[62952]: DEBUG nova.virt.hardware [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1429.399482] env[62952]: DEBUG nova.virt.hardware [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1429.400411] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-174ae53b-66c4-4ec6-aa83-8b4e3971c39d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.408241] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f79b05a-3db4-4714-b46c-511a814bcaf4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.421357] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:88:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3f499bc9-78da-46c1-9274-19edf26d31cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ae6d72db-0cb4-4fc3-a905-e42445320885', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1429.428642] env[62952]: DEBUG oslo.service.loopingcall [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1429.428865] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1429.429078] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e3bc348-7f24-4a5f-bfa5-933d0d327373 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.446871] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1429.446871] env[62952]: value = "task-1264099" [ 1429.446871] env[62952]: _type = "Task" [ 1429.446871] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1429.453792] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1264099, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1429.957054] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1264099, 'name': CreateVM_Task, 'duration_secs': 0.29225} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1429.957236] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1429.957926] env[62952]: DEBUG oslo_concurrency.lockutils [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5686e478-a71e-4391-b5e5-6826db862980" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1429.958110] env[62952]: DEBUG oslo_concurrency.lockutils [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5686e478-a71e-4391-b5e5-6826db862980" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1429.958554] env[62952]: DEBUG oslo_concurrency.lockutils [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5686e478-a71e-4391-b5e5-6826db862980" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1429.958813] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1c69cb5-e57f-4152-9849-873a9c3c1f4a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.963167] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1429.963167] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ef02a6-07f8-ff75-2773-adafa0dae712" [ 1429.963167] env[62952]: _type = "Task" [ 1429.963167] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1429.971636] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ef02a6-07f8-ff75-2773-adafa0dae712, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1430.103948] env[62952]: DEBUG nova.compute.manager [req-44b5d84c-3c88-460a-8c50-63a25c6934c0 req-3247e76b-eb7e-46ea-92f2-1e76c2769e5d service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Received event network-changed-ae6d72db-0cb4-4fc3-a905-e42445320885 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1430.104174] env[62952]: DEBUG nova.compute.manager [req-44b5d84c-3c88-460a-8c50-63a25c6934c0 req-3247e76b-eb7e-46ea-92f2-1e76c2769e5d service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Refreshing instance network info cache due to event network-changed-ae6d72db-0cb4-4fc3-a905-e42445320885. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1430.104395] env[62952]: DEBUG oslo_concurrency.lockutils [req-44b5d84c-3c88-460a-8c50-63a25c6934c0 req-3247e76b-eb7e-46ea-92f2-1e76c2769e5d service nova] Acquiring lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1430.104544] env[62952]: DEBUG oslo_concurrency.lockutils [req-44b5d84c-3c88-460a-8c50-63a25c6934c0 req-3247e76b-eb7e-46ea-92f2-1e76c2769e5d service nova] Acquired lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1430.104702] env[62952]: DEBUG nova.network.neutron [req-44b5d84c-3c88-460a-8c50-63a25c6934c0 req-3247e76b-eb7e-46ea-92f2-1e76c2769e5d service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Refreshing network info cache for port ae6d72db-0cb4-4fc3-a905-e42445320885 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1430.473627] env[62952]: DEBUG oslo_concurrency.lockutils [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5686e478-a71e-4391-b5e5-6826db862980" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1430.474013] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Processing image 5686e478-a71e-4391-b5e5-6826db862980 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1430.474197] env[62952]: DEBUG oslo_concurrency.lockutils [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5686e478-a71e-4391-b5e5-6826db862980/5686e478-a71e-4391-b5e5-6826db862980.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1430.474261] env[62952]: DEBUG oslo_concurrency.lockutils [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5686e478-a71e-4391-b5e5-6826db862980/5686e478-a71e-4391-b5e5-6826db862980.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1430.474442] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1430.474688] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8fbdf09a-ba61-48f4-9ad2-60f6a6026c03 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.482603] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1430.482773] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1430.483526] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51c36326-b0aa-496b-8be5-988962fac040 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.487980] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1430.487980] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e13346-2aa3-b368-96d3-503bc4df83b0" [ 1430.487980] env[62952]: _type = "Task" [ 1430.487980] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1430.494889] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52e13346-2aa3-b368-96d3-503bc4df83b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1430.800392] env[62952]: DEBUG nova.network.neutron [req-44b5d84c-3c88-460a-8c50-63a25c6934c0 req-3247e76b-eb7e-46ea-92f2-1e76c2769e5d service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Updated VIF entry in instance network info cache for port ae6d72db-0cb4-4fc3-a905-e42445320885. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1430.800742] env[62952]: DEBUG nova.network.neutron [req-44b5d84c-3c88-460a-8c50-63a25c6934c0 req-3247e76b-eb7e-46ea-92f2-1e76c2769e5d service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Updating instance_info_cache with network_info: [{"id": "ae6d72db-0cb4-4fc3-a905-e42445320885", "address": "fa:16:3e:a9:88:98", "network": {"id": "df16fd01-e3e5-47d2-8992-e42d8db894e0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1554441873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2f92c0076c14f15b5a4b8188f50db8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f499bc9-78da-46c1-9274-19edf26d31cb", "external-id": "nsx-vlan-transportzone-243", "segmentation_id": 243, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae6d72db-0c", "ovs_interfaceid": "ae6d72db-0cb4-4fc3-a905-e42445320885", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1430.997666] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Preparing fetch location {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1430.997917] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Fetch image to [datastore1] OSTACK_IMG_d9840db3-468b-4e6b-bc5d-f320cea82ff5/OSTACK_IMG_d9840db3-468b-4e6b-bc5d-f320cea82ff5.vmdk {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1430.998121] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Downloading stream optimized image 5686e478-a71e-4391-b5e5-6826db862980 to [datastore1] OSTACK_IMG_d9840db3-468b-4e6b-bc5d-f320cea82ff5/OSTACK_IMG_d9840db3-468b-4e6b-bc5d-f320cea82ff5.vmdk on the data store datastore1 as vApp {{(pid=62952) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1430.998297] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Downloading image file data 5686e478-a71e-4391-b5e5-6826db862980 to the ESX as VM named 'OSTACK_IMG_d9840db3-468b-4e6b-bc5d-f320cea82ff5' {{(pid=62952) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1431.063611] env[62952]: DEBUG oslo_vmware.rw_handles [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1431.063611] env[62952]: value = "resgroup-9" [ 1431.063611] env[62952]: _type = "ResourcePool" [ 1431.063611] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1431.063901] env[62952]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-ee7ea7e5-46fe-4dc3-a27e-90cb9e617c7e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.086092] env[62952]: DEBUG oslo_vmware.rw_handles [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lease: (returnval){ [ 1431.086092] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ac141e-21a2-5e6a-e93d-a26acdb082a1" [ 1431.086092] env[62952]: _type = "HttpNfcLease" [ 1431.086092] env[62952]: } obtained for vApp import into resource pool (val){ [ 1431.086092] env[62952]: value = "resgroup-9" [ 1431.086092] env[62952]: _type = "ResourcePool" [ 1431.086092] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1431.086092] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the lease: (returnval){ [ 1431.086092] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ac141e-21a2-5e6a-e93d-a26acdb082a1" [ 1431.086092] env[62952]: _type = "HttpNfcLease" [ 1431.086092] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1431.092237] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1431.092237] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ac141e-21a2-5e6a-e93d-a26acdb082a1" [ 1431.092237] env[62952]: _type = "HttpNfcLease" [ 1431.092237] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1431.304072] env[62952]: DEBUG oslo_concurrency.lockutils [req-44b5d84c-3c88-460a-8c50-63a25c6934c0 req-3247e76b-eb7e-46ea-92f2-1e76c2769e5d service nova] Releasing lock "refresh_cache-9f3466ae-f814-4831-b485-e8c5c28e02ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1431.594646] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1431.594646] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ac141e-21a2-5e6a-e93d-a26acdb082a1" [ 1431.594646] env[62952]: _type = "HttpNfcLease" [ 1431.594646] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1431.595078] env[62952]: DEBUG oslo_vmware.rw_handles [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1431.595078] env[62952]: value = "session[5259aeaa-aab6-02e3-be2d-7c693f3b3da4]52ac141e-21a2-5e6a-e93d-a26acdb082a1" [ 1431.595078] env[62952]: _type = "HttpNfcLease" [ 1431.595078] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1431.595622] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd02148-7d70-4eec-9527-cee1dc4bf304 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.602571] env[62952]: DEBUG oslo_vmware.rw_handles [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522f2d83-895e-1809-ecab-8f10360c2f56/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1431.602751] env[62952]: DEBUG oslo_vmware.rw_handles [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522f2d83-895e-1809-ecab-8f10360c2f56/disk-0.vmdk. {{(pid=62952) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1431.667513] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-537b4c32-ab2c-4b6c-89e3-1b889c52fadd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.717304] env[62952]: DEBUG oslo_vmware.rw_handles [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Completed reading data from the image iterator. {{(pid=62952) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1432.717738] env[62952]: DEBUG oslo_vmware.rw_handles [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522f2d83-895e-1809-ecab-8f10360c2f56/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1432.718444] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2503fe04-3898-4384-92c7-b9a3267aa574 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.725993] env[62952]: DEBUG oslo_vmware.rw_handles [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522f2d83-895e-1809-ecab-8f10360c2f56/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1432.726210] env[62952]: DEBUG oslo_vmware.rw_handles [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522f2d83-895e-1809-ecab-8f10360c2f56/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1432.726504] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-f2cc30ca-5c33-4a09-9b12-94cf3e5dc6b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.910499] env[62952]: DEBUG oslo_vmware.rw_handles [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522f2d83-895e-1809-ecab-8f10360c2f56/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1432.910758] env[62952]: INFO nova.virt.vmwareapi.images [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Downloaded image file data 5686e478-a71e-4391-b5e5-6826db862980 [ 1432.911661] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28eb148a-ffae-43d3-90a6-2723433d486f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.927096] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b1302294-2c68-40b1-97f3-f7f21d563136 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.956542] env[62952]: INFO nova.virt.vmwareapi.images [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] The imported VM was unregistered [ 1432.958988] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Caching image {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1432.959263] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Creating directory with path [datastore1] devstack-image-cache_base/5686e478-a71e-4391-b5e5-6826db862980 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1432.959583] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d75eaf8-204f-41f7-b189-c8150308380d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.968694] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Created directory with path [datastore1] devstack-image-cache_base/5686e478-a71e-4391-b5e5-6826db862980 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1432.968832] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_d9840db3-468b-4e6b-bc5d-f320cea82ff5/OSTACK_IMG_d9840db3-468b-4e6b-bc5d-f320cea82ff5.vmdk to [datastore1] devstack-image-cache_base/5686e478-a71e-4391-b5e5-6826db862980/5686e478-a71e-4391-b5e5-6826db862980.vmdk. {{(pid=62952) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1432.969067] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-41dddf1f-5f56-4fb0-8351-b1456b36cb9b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.975426] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1432.975426] env[62952]: value = "task-1264102" [ 1432.975426] env[62952]: _type = "Task" [ 1432.975426] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1432.983776] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264102, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1433.488408] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264102, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1433.988292] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264102, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1434.489233] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264102, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1434.989739] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264102, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1435.488559] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264102, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.348463} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1435.488834] env[62952]: INFO nova.virt.vmwareapi.ds_util [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_d9840db3-468b-4e6b-bc5d-f320cea82ff5/OSTACK_IMG_d9840db3-468b-4e6b-bc5d-f320cea82ff5.vmdk to [datastore1] devstack-image-cache_base/5686e478-a71e-4391-b5e5-6826db862980/5686e478-a71e-4391-b5e5-6826db862980.vmdk. [ 1435.489035] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Cleaning up location [datastore1] OSTACK_IMG_d9840db3-468b-4e6b-bc5d-f320cea82ff5 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1435.489211] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_d9840db3-468b-4e6b-bc5d-f320cea82ff5 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1435.489491] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0bc42af0-df10-4d25-9763-b7e636aceeb4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.496057] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1435.496057] env[62952]: value = "task-1264103" [ 1435.496057] env[62952]: _type = "Task" [ 1435.496057] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1435.503322] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264103, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1436.006395] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264103, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.066984} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1436.006765] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1436.006862] env[62952]: DEBUG oslo_concurrency.lockutils [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5686e478-a71e-4391-b5e5-6826db862980/5686e478-a71e-4391-b5e5-6826db862980.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1436.007101] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5686e478-a71e-4391-b5e5-6826db862980/5686e478-a71e-4391-b5e5-6826db862980.vmdk to [datastore1] 9f3466ae-f814-4831-b485-e8c5c28e02ea/9f3466ae-f814-4831-b485-e8c5c28e02ea.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1436.007364] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-26d02c33-e814-4992-b6cc-0e1e180480ac {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.013630] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1436.013630] env[62952]: value = "task-1264104" [ 1436.013630] env[62952]: _type = "Task" [ 1436.013630] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1436.020671] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264104, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1436.525886] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264104, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1437.026315] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264104, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1437.528353] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264104, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1438.028772] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264104, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1438.527792] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264104, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.077213} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1438.528038] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5686e478-a71e-4391-b5e5-6826db862980/5686e478-a71e-4391-b5e5-6826db862980.vmdk to [datastore1] 9f3466ae-f814-4831-b485-e8c5c28e02ea/9f3466ae-f814-4831-b485-e8c5c28e02ea.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1438.528778] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1db4a8f3-de2f-47ea-afcc-b4f5cc02d52d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1438.549691] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 9f3466ae-f814-4831-b485-e8c5c28e02ea/9f3466ae-f814-4831-b485-e8c5c28e02ea.vmdk or device None with type streamOptimized {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1438.549925] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2725c44-c2e0-477a-9f42-2ba43c5ccfdc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1438.568865] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1438.568865] env[62952]: value = "task-1264105" [ 1438.568865] env[62952]: _type = "Task" [ 1438.568865] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1438.576036] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264105, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1439.078172] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264105, 'name': ReconfigVM_Task, 'duration_secs': 0.294766} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1439.078464] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 9f3466ae-f814-4831-b485-e8c5c28e02ea/9f3466ae-f814-4831-b485-e8c5c28e02ea.vmdk or device None with type streamOptimized {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1439.079094] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c1fc2ec4-9175-4525-a2b9-b9021cc2617d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.085899] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1439.085899] env[62952]: value = "task-1264106" [ 1439.085899] env[62952]: _type = "Task" [ 1439.085899] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1439.094103] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264106, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1439.595523] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264106, 'name': Rename_Task, 'duration_secs': 0.174667} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1439.595773] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1439.596016] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1f466f2-73f7-40d9-b296-d93cdbb25892 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.602350] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1439.602350] env[62952]: value = "task-1264107" [ 1439.602350] env[62952]: _type = "Task" [ 1439.602350] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1439.609319] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264107, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1440.112778] env[62952]: DEBUG oslo_vmware.api [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264107, 'name': PowerOnVM_Task, 'duration_secs': 0.468702} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1440.113072] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1440.211702] env[62952]: DEBUG nova.compute.manager [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1440.212674] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f826f47-c710-4c70-9610-87971c4f1487 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.728956] env[62952]: DEBUG oslo_concurrency.lockutils [None req-45d35a42-8e43-4661-bb2e-842f3defc290 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "9f3466ae-f814-4831-b485-e8c5c28e02ea" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.186s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1442.022327] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "9f3466ae-f814-4831-b485-e8c5c28e02ea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1442.022718] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "9f3466ae-f814-4831-b485-e8c5c28e02ea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1442.022866] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "9f3466ae-f814-4831-b485-e8c5c28e02ea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1442.023072] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "9f3466ae-f814-4831-b485-e8c5c28e02ea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1442.023252] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "9f3466ae-f814-4831-b485-e8c5c28e02ea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1442.025592] env[62952]: INFO nova.compute.manager [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Terminating instance [ 1442.027417] env[62952]: DEBUG nova.compute.manager [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1442.027615] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1442.028476] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c86eb5c4-0782-4b46-ab42-d80e62f8a152 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.036682] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1442.036897] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9565c8de-3912-422e-bf0a-37f86831e50d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.042418] env[62952]: DEBUG oslo_vmware.api [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1442.042418] env[62952]: value = "task-1264108" [ 1442.042418] env[62952]: _type = "Task" [ 1442.042418] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1442.049939] env[62952]: DEBUG oslo_vmware.api [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264108, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1442.551778] env[62952]: DEBUG oslo_vmware.api [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264108, 'name': PowerOffVM_Task, 'duration_secs': 0.171498} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1442.552071] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1442.552222] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1442.552458] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-727d9f56-e71f-4e80-aa08-2b357e38c086 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.611641] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1442.611933] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1442.612116] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Deleting the datastore file [datastore1] 9f3466ae-f814-4831-b485-e8c5c28e02ea {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1442.612324] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-16c067af-45a5-46e2-8a61-83f6fc6df683 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.619352] env[62952]: DEBUG oslo_vmware.api [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for the task: (returnval){ [ 1442.619352] env[62952]: value = "task-1264110" [ 1442.619352] env[62952]: _type = "Task" [ 1442.619352] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1442.626571] env[62952]: DEBUG oslo_vmware.api [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264110, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1443.128835] env[62952]: DEBUG oslo_vmware.api [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Task: {'id': task-1264110, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.11563} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1443.129235] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1443.129312] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1443.129441] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1443.129626] env[62952]: INFO nova.compute.manager [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1443.129869] env[62952]: DEBUG oslo.service.loopingcall [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1443.130072] env[62952]: DEBUG nova.compute.manager [-] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1443.130169] env[62952]: DEBUG nova.network.neutron [-] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1443.555073] env[62952]: DEBUG nova.compute.manager [req-1270717b-d8a6-4b92-8f9a-55dc1c9c4db6 req-f0452c8e-4bf0-4add-a6f0-56d14062d3b1 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Received event network-vif-deleted-ae6d72db-0cb4-4fc3-a905-e42445320885 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1443.555225] env[62952]: INFO nova.compute.manager [req-1270717b-d8a6-4b92-8f9a-55dc1c9c4db6 req-f0452c8e-4bf0-4add-a6f0-56d14062d3b1 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Neutron deleted interface ae6d72db-0cb4-4fc3-a905-e42445320885; detaching it from the instance and deleting it from the info cache [ 1443.555426] env[62952]: DEBUG nova.network.neutron [req-1270717b-d8a6-4b92-8f9a-55dc1c9c4db6 req-f0452c8e-4bf0-4add-a6f0-56d14062d3b1 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1444.034595] env[62952]: DEBUG nova.network.neutron [-] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1444.058225] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c6678e2-bd10-41e5-9c0f-3d3656d448b4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1444.067429] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02caab5-2274-4601-ade6-fe0c06990cab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1444.091122] env[62952]: DEBUG nova.compute.manager [req-1270717b-d8a6-4b92-8f9a-55dc1c9c4db6 req-f0452c8e-4bf0-4add-a6f0-56d14062d3b1 service nova] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Detach interface failed, port_id=ae6d72db-0cb4-4fc3-a905-e42445320885, reason: Instance 9f3466ae-f814-4831-b485-e8c5c28e02ea could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1444.537377] env[62952]: INFO nova.compute.manager [-] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Took 1.41 seconds to deallocate network for instance. [ 1445.044314] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1445.044588] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1445.044816] env[62952]: DEBUG nova.objects.instance [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lazy-loading 'resources' on Instance uuid 9f3466ae-f814-4831-b485-e8c5c28e02ea {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1445.565110] env[62952]: DEBUG nova.scheduler.client.report [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Refreshing inventories for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1445.579542] env[62952]: DEBUG nova.scheduler.client.report [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Updating ProviderTree inventory for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1445.579810] env[62952]: DEBUG nova.compute.provider_tree [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Updating inventory in ProviderTree for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1445.589455] env[62952]: DEBUG nova.scheduler.client.report [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Refreshing aggregate associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1445.605817] env[62952]: DEBUG nova.scheduler.client.report [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Refreshing trait associations for resource provider 615127b5-dc82-4491-9c8f-4e6d0caa0690, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1445.638912] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc509fe-bbfa-4abb-8ccd-1b1a68f95f47 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1445.647135] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e747a0-e285-4f39-a25d-b104bf437188 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1445.675629] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-507711a8-b1c2-48e7-b941-a6c3462fa772 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1445.682253] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25f72028-cd25-48a4-a905-78037f35f0ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1445.696017] env[62952]: DEBUG nova.compute.provider_tree [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1446.199155] env[62952]: DEBUG nova.scheduler.client.report [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1446.704391] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.660s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1446.724637] env[62952]: INFO nova.scheduler.client.report [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Deleted allocations for instance 9f3466ae-f814-4831-b485-e8c5c28e02ea [ 1447.232796] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b25d2d71-16f8-4b80-8df3-58fb00f6a2d8 tempest-ServerActionsTestOtherB-1091171981 tempest-ServerActionsTestOtherB-1091171981-project-member] Lock "9f3466ae-f814-4831-b485-e8c5c28e02ea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.210s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1447.272607] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1447.272789] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Cleaning up deleted instances {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1447.778019] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] There are 10 instances to clean {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1447.778399] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 9f3466ae-f814-4831-b485-e8c5c28e02ea] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1448.231770] env[62952]: DEBUG oslo_concurrency.lockutils [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1448.232182] env[62952]: DEBUG oslo_concurrency.lockutils [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1448.281166] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: b5da3312-a52c-4941-9c3c-247f4d93fbbf] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1448.735882] env[62952]: DEBUG nova.compute.utils [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1448.785146] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 34edb671-8d37-4c08-b8ca-3b4b89b345b9] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1449.238754] env[62952]: DEBUG oslo_concurrency.lockutils [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1449.287883] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: cf523cf4-028c-4a24-9822-df384974458e] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1449.792034] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: c5d21282-1a4b-4998-8170-ee5423b958ad] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1450.297134] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: b01ea5c5-1d49-47e7-ada2-acbacbe5abdb] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1450.299566] env[62952]: DEBUG oslo_concurrency.lockutils [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1450.299936] env[62952]: DEBUG oslo_concurrency.lockutils [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1450.300284] env[62952]: INFO nova.compute.manager [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Attaching volume 0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5 to /dev/sdb [ 1450.335806] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0689253-e193-4739-ad13-eda8a114323d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1450.343619] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b04e930-517e-42b3-9b6b-1566eae8e24b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1450.356676] env[62952]: DEBUG nova.virt.block_device [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Updating existing volume attachment record: c8c201ad-dc67-4702-94a0-ede00f327913 {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1450.804360] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: f92ef782-fd64-47f0-8145-9261a1c3005c] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1451.307884] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 43d79d26-4350-4329-bffc-cedfada40ee0] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1451.811364] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 42793459-89e6-4941-9b20-cbe52a241c6d] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1452.314324] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: dbca4369-1ed3-493a-9847-9fa1e4293475] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1454.272658] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1454.272943] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1454.272986] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1454.813074] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "refresh_cache-7bf9931c-18a3-4bb5-99d2-a45263c6b236" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1454.813074] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquired lock "refresh_cache-7bf9931c-18a3-4bb5-99d2-a45263c6b236" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1454.813074] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Forcefully refreshing network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1454.813074] env[62952]: DEBUG nova.objects.instance [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lazy-loading 'info_cache' on Instance uuid 7bf9931c-18a3-4bb5-99d2-a45263c6b236 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1454.911848] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1454.912142] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272008', 'volume_id': '0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5', 'name': 'volume-0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7bf9931c-18a3-4bb5-99d2-a45263c6b236', 'attached_at': '', 'detached_at': '', 'volume_id': '0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5', 'serial': '0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1454.913053] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a017e685-e54c-4f4c-b720-3c9c054c265b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1454.930584] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33bc2198-cb95-4960-ac26-19fcfde4a0de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1454.957221] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] volume-0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5/volume-0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1454.957583] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb568ddc-3bd0-43e7-96df-be5eae314ff8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1454.977862] env[62952]: DEBUG oslo_vmware.api [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1454.977862] env[62952]: value = "task-1264114" [ 1454.977862] env[62952]: _type = "Task" [ 1454.977862] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1454.985763] env[62952]: DEBUG oslo_vmware.api [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264114, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1455.487840] env[62952]: DEBUG oslo_vmware.api [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264114, 'name': ReconfigVM_Task, 'duration_secs': 0.321288} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1455.488321] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Reconfigured VM instance instance-00000070 to attach disk [datastore2] volume-0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5/volume-0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1455.492948] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77bacddd-f514-4ccc-9fc3-5d2a69183c1e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1455.507981] env[62952]: DEBUG oslo_vmware.api [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1455.507981] env[62952]: value = "task-1264115" [ 1455.507981] env[62952]: _type = "Task" [ 1455.507981] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1455.515973] env[62952]: DEBUG oslo_vmware.api [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264115, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1456.019104] env[62952]: DEBUG oslo_vmware.api [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264115, 'name': ReconfigVM_Task, 'duration_secs': 0.131179} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1456.019454] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272008', 'volume_id': '0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5', 'name': 'volume-0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7bf9931c-18a3-4bb5-99d2-a45263c6b236', 'attached_at': '', 'detached_at': '', 'volume_id': '0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5', 'serial': '0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1456.548683] env[62952]: DEBUG nova.network.neutron [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Updating instance_info_cache with network_info: [{"id": "6c006fd9-5eb1-4220-9160-af6ec0ca7156", "address": "fa:16:3e:3c:9b:34", "network": {"id": "8e167a8b-2628-43e1-8418-e673064745b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-232555460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f725be1feb8a44478d297b0be58c59ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c006fd9-5e", "ovs_interfaceid": "6c006fd9-5eb1-4220-9160-af6ec0ca7156", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1457.051921] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Releasing lock "refresh_cache-7bf9931c-18a3-4bb5-99d2-a45263c6b236" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1457.052093] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Updated the network info_cache for instance {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1457.052317] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1457.052475] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1457.052619] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1457.052818] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1457.053010] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1457.053157] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1457.053307] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1457.055402] env[62952]: DEBUG nova.objects.instance [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lazy-loading 'flavor' on Instance uuid 7bf9931c-18a3-4bb5-99d2-a45263c6b236 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1457.555882] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1457.556190] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1457.556302] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1457.556459] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1457.557408] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba5b449-9822-4896-a131-8ce676edd600 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1457.562657] env[62952]: DEBUG oslo_concurrency.lockutils [None req-642431b3-0e4a-4186-b718-98f89007ef25 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.263s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1457.567677] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc56f923-6d42-4465-aaa8-b4d74f8cc00e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1457.582190] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594de86e-0f96-43f2-a497-e98d0d97b2ba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1457.588776] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd7e1f8b-b5f1-4adc-91e1-775e6a9e6b43 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1457.616583] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181470MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1457.616725] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1457.616908] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1457.759608] env[62952]: DEBUG oslo_concurrency.lockutils [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1457.759883] env[62952]: DEBUG oslo_concurrency.lockutils [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1458.263586] env[62952]: INFO nova.compute.manager [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Detaching volume 0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5 [ 1458.293300] env[62952]: INFO nova.virt.block_device [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Attempting to driver detach volume 0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5 from mountpoint /dev/sdb [ 1458.293887] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1458.293887] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272008', 'volume_id': '0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5', 'name': 'volume-0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7bf9931c-18a3-4bb5-99d2-a45263c6b236', 'attached_at': '', 'detached_at': '', 'volume_id': '0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5', 'serial': '0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1458.294606] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b96fbb20-7e58-4ff0-852b-7d0a43cca49f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1458.315470] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-231bddab-1332-45cb-8c7a-9f23e047935a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1458.321852] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eea6a121-1721-49a7-a5e6-c300966a7bd1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1458.342115] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c9cbd0-3c0c-407e-9bfb-dd2f94fddcc8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1458.355843] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] The volume has not been displaced from its original location: [datastore2] volume-0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5/volume-0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1458.360885] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Reconfiguring VM instance instance-00000070 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1458.361184] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-660da0a6-0d53-4764-8ae9-ca26ca3d29c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1458.377599] env[62952]: DEBUG oslo_vmware.api [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1458.377599] env[62952]: value = "task-1264116" [ 1458.377599] env[62952]: _type = "Task" [ 1458.377599] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1458.384714] env[62952]: DEBUG oslo_vmware.api [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264116, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1458.756385] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Instance 7bf9931c-18a3-4bb5-99d2-a45263c6b236 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1458.756656] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1458.756732] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1458.784871] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519daa10-2cbf-4e82-be02-c0480417979b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1458.792545] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca996ed1-3ffc-4a7d-884e-4a7034cb89bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1458.820790] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743ed669-5d59-4b48-b5d7-437e51ac82d3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1458.827614] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a22d6d7-9278-4e76-8cb0-0f5745cd56f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1458.840907] env[62952]: DEBUG nova.compute.provider_tree [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1458.886466] env[62952]: DEBUG oslo_vmware.api [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264116, 'name': ReconfigVM_Task, 'duration_secs': 0.209881} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1458.886688] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Reconfigured VM instance instance-00000070 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1458.891160] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-671acf2a-310e-4667-996e-fb6db24e2460 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1458.904869] env[62952]: DEBUG oslo_vmware.api [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1458.904869] env[62952]: value = "task-1264117" [ 1458.904869] env[62952]: _type = "Task" [ 1458.904869] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1458.911947] env[62952]: DEBUG oslo_vmware.api [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264117, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1459.344256] env[62952]: DEBUG nova.scheduler.client.report [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1459.415474] env[62952]: DEBUG oslo_vmware.api [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264117, 'name': ReconfigVM_Task, 'duration_secs': 0.129575} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1459.415731] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272008', 'volume_id': '0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5', 'name': 'volume-0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7bf9931c-18a3-4bb5-99d2-a45263c6b236', 'attached_at': '', 'detached_at': '', 'volume_id': '0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5', 'serial': '0dcaaa2b-5ba4-4f00-a006-fcbd4b733bc5'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1459.849327] env[62952]: DEBUG nova.compute.resource_tracker [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1459.849700] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.233s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1459.849825] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1459.849924] env[62952]: DEBUG nova.compute.manager [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Cleaning up deleted instances with incomplete migration {{(pid=62952) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1459.958641] env[62952]: DEBUG nova.objects.instance [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lazy-loading 'flavor' on Instance uuid 7bf9931c-18a3-4bb5-99d2-a45263c6b236 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1460.966546] env[62952]: DEBUG oslo_concurrency.lockutils [None req-34e7ea61-3989-4544-90b4-b7582f177b31 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.206s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1461.571436] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1461.571684] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1461.992541] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1461.992912] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1461.993137] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1461.993337] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1461.993517] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1461.995772] env[62952]: INFO nova.compute.manager [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Terminating instance [ 1461.997479] env[62952]: DEBUG nova.compute.manager [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1461.997673] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1461.998555] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd8b6a7-855b-4c9a-b5c3-ffdb8f5ab0bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1462.006372] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1462.006620] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb796067-4e06-46a9-b71d-acb5fcfd8258 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1462.013174] env[62952]: DEBUG oslo_vmware.api [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1462.013174] env[62952]: value = "task-1264118" [ 1462.013174] env[62952]: _type = "Task" [ 1462.013174] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1462.020821] env[62952]: DEBUG oslo_vmware.api [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264118, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1462.523501] env[62952]: DEBUG oslo_vmware.api [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264118, 'name': PowerOffVM_Task, 'duration_secs': 0.169235} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1462.523772] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1462.523949] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1462.524204] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9fbf2007-28d0-4d59-adf3-1c6c27763a69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1462.583584] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1462.583827] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1462.584025] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Deleting the datastore file [datastore1] 7bf9931c-18a3-4bb5-99d2-a45263c6b236 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1462.584295] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ce14a697-0281-443c-9d8d-dd42250b4659 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1462.590083] env[62952]: DEBUG oslo_vmware.api [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for the task: (returnval){ [ 1462.590083] env[62952]: value = "task-1264120" [ 1462.590083] env[62952]: _type = "Task" [ 1462.590083] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1462.597291] env[62952]: DEBUG oslo_vmware.api [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264120, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1463.101070] env[62952]: DEBUG oslo_vmware.api [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Task: {'id': task-1264120, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.125572} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1463.101459] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1463.101566] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1463.101769] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1463.101996] env[62952]: INFO nova.compute.manager [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1463.102288] env[62952]: DEBUG oslo.service.loopingcall [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1463.102531] env[62952]: DEBUG nova.compute.manager [-] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1463.102662] env[62952]: DEBUG nova.network.neutron [-] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1463.560699] env[62952]: DEBUG nova.compute.manager [req-a7832b10-bac1-4545-8a65-2475638f9e78 req-ff99725e-84c0-4d58-a1ad-8d5d4e4a92d0 service nova] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Received event network-vif-deleted-6c006fd9-5eb1-4220-9160-af6ec0ca7156 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1463.560981] env[62952]: INFO nova.compute.manager [req-a7832b10-bac1-4545-8a65-2475638f9e78 req-ff99725e-84c0-4d58-a1ad-8d5d4e4a92d0 service nova] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Neutron deleted interface 6c006fd9-5eb1-4220-9160-af6ec0ca7156; detaching it from the instance and deleting it from the info cache [ 1463.561148] env[62952]: DEBUG nova.network.neutron [req-a7832b10-bac1-4545-8a65-2475638f9e78 req-ff99725e-84c0-4d58-a1ad-8d5d4e4a92d0 service nova] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1464.038279] env[62952]: DEBUG nova.network.neutron [-] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1464.064345] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fb393794-558d-4c60-994c-c0d611233d91 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1464.073622] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6543bf39-13a7-455c-af8a-444363d18f6f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1464.097013] env[62952]: DEBUG nova.compute.manager [req-a7832b10-bac1-4545-8a65-2475638f9e78 req-ff99725e-84c0-4d58-a1ad-8d5d4e4a92d0 service nova] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Detach interface failed, port_id=6c006fd9-5eb1-4220-9160-af6ec0ca7156, reason: Instance 7bf9931c-18a3-4bb5-99d2-a45263c6b236 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1464.541059] env[62952]: INFO nova.compute.manager [-] [instance: 7bf9931c-18a3-4bb5-99d2-a45263c6b236] Took 1.44 seconds to deallocate network for instance. [ 1465.047776] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1465.048080] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1465.048341] env[62952]: DEBUG nova.objects.instance [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lazy-loading 'resources' on Instance uuid 7bf9931c-18a3-4bb5-99d2-a45263c6b236 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1465.272020] env[62952]: DEBUG oslo_service.periodic_task [None req-aa44a270-0550-4147-82df-9e971d006017 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1465.584586] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-886a0aa7-08f8-4e1f-ad01-918cff2fd259 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.592431] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba96d6c8-7a47-4589-b67a-a4b7aaec931f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.624265] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7295c83b-fa2e-4860-a37a-6e7dd3f4a518 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.631523] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e08e48-1aa2-40aa-b920-cab1d2fe7e1c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.644206] env[62952]: DEBUG nova.compute.provider_tree [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Inventory has not changed in ProviderTree for provider: 615127b5-dc82-4491-9c8f-4e6d0caa0690 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1466.147740] env[62952]: DEBUG nova.scheduler.client.report [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Inventory has not changed for provider 615127b5-dc82-4491-9c8f-4e6d0caa0690 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1466.653283] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.605s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1466.676980] env[62952]: INFO nova.scheduler.client.report [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Deleted allocations for instance 7bf9931c-18a3-4bb5-99d2-a45263c6b236 [ 1467.184556] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6aa8f2e8-52a5-4909-94c3-1a9b56e73500 tempest-AttachVolumeNegativeTest-1921176247 tempest-AttachVolumeNegativeTest-1921176247-project-member] Lock "7bf9931c-18a3-4bb5-99d2-a45263c6b236" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.192s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}